var/home/core/zuul-output/0000755000175000017500000000000015114271350014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114316106015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006770645015114316077017721 0ustar rootrootDec 04 11:42:54 crc systemd[1]: Starting Kubernetes Kubelet... Dec 04 11:42:54 crc restorecon[4745]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:54 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 04 11:42:55 crc restorecon[4745]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 04 11:42:56 crc kubenswrapper[4979]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 11:42:56 crc kubenswrapper[4979]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 04 11:42:56 crc kubenswrapper[4979]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 11:42:56 crc kubenswrapper[4979]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 11:42:56 crc kubenswrapper[4979]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 04 11:42:56 crc kubenswrapper[4979]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.039881 4979 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044433 4979 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044465 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044473 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044481 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044489 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044498 4979 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044506 4979 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044513 4979 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044519 4979 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044525 4979 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044531 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044536 4979 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044541 4979 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044547 4979 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044554 4979 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044559 4979 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044566 4979 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044573 4979 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044579 4979 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044584 4979 feature_gate.go:330] unrecognized feature gate: Example Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044589 4979 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044594 4979 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044599 4979 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044604 4979 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044611 4979 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044616 4979 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044620 4979 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044625 4979 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044630 4979 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044636 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044641 4979 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044647 4979 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044653 4979 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044661 4979 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044667 4979 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044674 4979 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044681 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044689 4979 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044695 4979 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044701 4979 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044707 4979 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044714 4979 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044719 4979 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044725 4979 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044732 4979 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044738 4979 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044744 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044750 4979 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044755 4979 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044760 4979 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044765 4979 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044770 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044775 4979 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044780 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044785 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044790 4979 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044795 4979 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044803 4979 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044810 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044817 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044824 4979 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044828 4979 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044835 4979 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044841 4979 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044847 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044854 4979 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044859 4979 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044864 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044870 4979 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044876 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.044881 4979 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045285 4979 flags.go:64] FLAG: --address="0.0.0.0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045329 4979 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045345 4979 flags.go:64] FLAG: --anonymous-auth="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045353 4979 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045363 4979 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045369 4979 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045377 4979 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045384 4979 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045390 4979 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045396 4979 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045402 4979 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045409 4979 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045416 4979 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045425 4979 flags.go:64] FLAG: --cgroup-root="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045434 4979 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045442 4979 flags.go:64] FLAG: --client-ca-file="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045450 4979 flags.go:64] FLAG: --cloud-config="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045458 4979 flags.go:64] FLAG: --cloud-provider="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045464 4979 flags.go:64] FLAG: --cluster-dns="[]" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045475 4979 flags.go:64] FLAG: --cluster-domain="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045482 4979 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045489 4979 flags.go:64] FLAG: --config-dir="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045496 4979 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045504 4979 flags.go:64] FLAG: --container-log-max-files="5" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045513 4979 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045522 4979 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045530 4979 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045538 4979 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045545 4979 flags.go:64] FLAG: --contention-profiling="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045552 4979 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045559 4979 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045565 4979 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045571 4979 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045578 4979 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045585 4979 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045592 4979 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045600 4979 flags.go:64] FLAG: --enable-load-reader="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045607 4979 flags.go:64] FLAG: --enable-server="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045614 4979 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045622 4979 flags.go:64] FLAG: --event-burst="100" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045630 4979 flags.go:64] FLAG: --event-qps="50" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045637 4979 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045644 4979 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045651 4979 flags.go:64] FLAG: --eviction-hard="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045661 4979 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045669 4979 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045675 4979 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045681 4979 flags.go:64] FLAG: --eviction-soft="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045687 4979 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045693 4979 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045700 4979 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045706 4979 flags.go:64] FLAG: --experimental-mounter-path="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045712 4979 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045718 4979 flags.go:64] FLAG: --fail-swap-on="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045723 4979 flags.go:64] FLAG: --feature-gates="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045730 4979 flags.go:64] FLAG: --file-check-frequency="20s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045738 4979 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045745 4979 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045753 4979 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045759 4979 flags.go:64] FLAG: --healthz-port="10248" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045766 4979 flags.go:64] FLAG: --help="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045774 4979 flags.go:64] FLAG: --hostname-override="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045780 4979 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045787 4979 flags.go:64] FLAG: --http-check-frequency="20s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045793 4979 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045799 4979 flags.go:64] FLAG: --image-credential-provider-config="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045805 4979 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045810 4979 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045816 4979 flags.go:64] FLAG: --image-service-endpoint="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045822 4979 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045829 4979 flags.go:64] FLAG: --kube-api-burst="100" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045836 4979 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045842 4979 flags.go:64] FLAG: --kube-api-qps="50" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045848 4979 flags.go:64] FLAG: --kube-reserved="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045853 4979 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045859 4979 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045865 4979 flags.go:64] FLAG: --kubelet-cgroups="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045871 4979 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045877 4979 flags.go:64] FLAG: --lock-file="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045882 4979 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045888 4979 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045894 4979 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045905 4979 flags.go:64] FLAG: --log-json-split-stream="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045912 4979 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045919 4979 flags.go:64] FLAG: --log-text-split-stream="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045926 4979 flags.go:64] FLAG: --logging-format="text" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045933 4979 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045939 4979 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045944 4979 flags.go:64] FLAG: --manifest-url="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045950 4979 flags.go:64] FLAG: --manifest-url-header="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045957 4979 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045963 4979 flags.go:64] FLAG: --max-open-files="1000000" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045970 4979 flags.go:64] FLAG: --max-pods="110" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045977 4979 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045983 4979 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045990 4979 flags.go:64] FLAG: --memory-manager-policy="None" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.045997 4979 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046005 4979 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046011 4979 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046017 4979 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046030 4979 flags.go:64] FLAG: --node-status-max-images="50" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046036 4979 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046042 4979 flags.go:64] FLAG: --oom-score-adj="-999" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046048 4979 flags.go:64] FLAG: --pod-cidr="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046053 4979 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046065 4979 flags.go:64] FLAG: --pod-manifest-path="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046073 4979 flags.go:64] FLAG: --pod-max-pids="-1" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046080 4979 flags.go:64] FLAG: --pods-per-core="0" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046086 4979 flags.go:64] FLAG: --port="10250" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046092 4979 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046098 4979 flags.go:64] FLAG: --provider-id="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046105 4979 flags.go:64] FLAG: --qos-reserved="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046115 4979 flags.go:64] FLAG: --read-only-port="10255" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046130 4979 flags.go:64] FLAG: --register-node="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046137 4979 flags.go:64] FLAG: --register-schedulable="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046147 4979 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046159 4979 flags.go:64] FLAG: --registry-burst="10" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046166 4979 flags.go:64] FLAG: --registry-qps="5" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046174 4979 flags.go:64] FLAG: --reserved-cpus="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046181 4979 flags.go:64] FLAG: --reserved-memory="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046190 4979 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046198 4979 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046204 4979 flags.go:64] FLAG: --rotate-certificates="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046210 4979 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046217 4979 flags.go:64] FLAG: --runonce="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046224 4979 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046230 4979 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046237 4979 flags.go:64] FLAG: --seccomp-default="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046243 4979 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046250 4979 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046256 4979 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046262 4979 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046267 4979 flags.go:64] FLAG: --storage-driver-password="root" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046274 4979 flags.go:64] FLAG: --storage-driver-secure="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046282 4979 flags.go:64] FLAG: --storage-driver-table="stats" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046292 4979 flags.go:64] FLAG: --storage-driver-user="root" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046332 4979 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046341 4979 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046349 4979 flags.go:64] FLAG: --system-cgroups="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046357 4979 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046368 4979 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046374 4979 flags.go:64] FLAG: --tls-cert-file="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046379 4979 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046388 4979 flags.go:64] FLAG: --tls-min-version="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046394 4979 flags.go:64] FLAG: --tls-private-key-file="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046401 4979 flags.go:64] FLAG: --topology-manager-policy="none" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046409 4979 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046417 4979 flags.go:64] FLAG: --topology-manager-scope="container" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046425 4979 flags.go:64] FLAG: --v="2" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046435 4979 flags.go:64] FLAG: --version="false" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046444 4979 flags.go:64] FLAG: --vmodule="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046453 4979 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.046461 4979 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046620 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046631 4979 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046637 4979 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046642 4979 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046647 4979 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046652 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046657 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046665 4979 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046672 4979 feature_gate.go:330] unrecognized feature gate: Example Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046680 4979 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046686 4979 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046693 4979 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046699 4979 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046705 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046713 4979 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046722 4979 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046730 4979 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046738 4979 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046746 4979 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046753 4979 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046761 4979 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046770 4979 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046779 4979 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046787 4979 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046795 4979 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046801 4979 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046809 4979 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046815 4979 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046824 4979 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046831 4979 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046836 4979 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046841 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046847 4979 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046852 4979 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046857 4979 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046862 4979 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046867 4979 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046908 4979 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046924 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046932 4979 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046940 4979 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046946 4979 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046952 4979 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046957 4979 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046962 4979 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046967 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046972 4979 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046977 4979 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046982 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046987 4979 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046991 4979 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.046996 4979 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047004 4979 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047010 4979 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047017 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047022 4979 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047027 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047033 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047040 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047045 4979 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047050 4979 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047055 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047060 4979 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047066 4979 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047073 4979 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047079 4979 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047084 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047090 4979 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047096 4979 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047102 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.047107 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.047422 4979 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.057895 4979 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.057973 4979 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058084 4979 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058110 4979 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058116 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058121 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058126 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058131 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058135 4979 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058142 4979 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058147 4979 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058152 4979 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058157 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058161 4979 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058183 4979 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058187 4979 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058190 4979 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058194 4979 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058198 4979 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058201 4979 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058205 4979 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058208 4979 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058212 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058215 4979 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058219 4979 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058222 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058226 4979 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058229 4979 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058234 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058238 4979 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058243 4979 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058262 4979 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058267 4979 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058271 4979 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058275 4979 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058280 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058284 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058288 4979 feature_gate.go:330] unrecognized feature gate: Example Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058292 4979 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058331 4979 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058339 4979 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058344 4979 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058348 4979 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058351 4979 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058355 4979 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058366 4979 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058370 4979 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058374 4979 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058378 4979 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058382 4979 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058386 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058407 4979 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058411 4979 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058416 4979 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058420 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058424 4979 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058428 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058431 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058435 4979 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058438 4979 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058443 4979 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058449 4979 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058452 4979 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058456 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058459 4979 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058463 4979 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058483 4979 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058489 4979 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058493 4979 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058496 4979 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058500 4979 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058503 4979 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058507 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.058513 4979 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058658 4979 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058666 4979 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058670 4979 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058674 4979 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058678 4979 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058683 4979 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058687 4979 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058692 4979 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058697 4979 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058716 4979 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058720 4979 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058724 4979 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058728 4979 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058732 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058735 4979 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058739 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058743 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058747 4979 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058752 4979 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058756 4979 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058760 4979 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058764 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058767 4979 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058771 4979 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058774 4979 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058793 4979 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058797 4979 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058801 4979 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058805 4979 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058808 4979 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058812 4979 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058815 4979 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058819 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058822 4979 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058826 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058829 4979 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058833 4979 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058837 4979 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058842 4979 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058846 4979 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058850 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058854 4979 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.058872 4979 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059000 4979 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059113 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059119 4979 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059124 4979 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059128 4979 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059149 4979 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059153 4979 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059157 4979 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059160 4979 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059165 4979 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.059168 4979 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060872 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060914 4979 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060919 4979 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060923 4979 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060927 4979 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060933 4979 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060937 4979 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060941 4979 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060945 4979 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060949 4979 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060953 4979 feature_gate.go:330] unrecognized feature gate: Example Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060957 4979 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.060992 4979 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.061001 4979 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.061009 4979 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.061017 4979 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.061024 4979 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.061033 4979 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.061325 4979 server.go:940] "Client rotation is on, will bootstrap in background" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.064668 4979 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.064767 4979 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.065372 4979 server.go:997] "Starting client certificate rotation" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.065406 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.065605 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-28 08:16:19.554440582 +0000 UTC Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.065757 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.074224 4979 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.075950 4979 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.078056 4979 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.084162 4979 log.go:25] "Validated CRI v1 runtime API" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.105793 4979 log.go:25] "Validated CRI v1 image API" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.108053 4979 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.112036 4979 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-04-11-38-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.112065 4979 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.128004 4979 manager.go:217] Machine: {Timestamp:2025-12-04 11:42:56.12400071 +0000 UTC m=+0.398296524 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:eae98eec-36de-419b-b26f-0e75001a46a0 BootID:7408354e-3656-48ab-94ea-2745a060a897 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:81:6e:7c Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:81:6e:7c Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9c:6b:a1 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:26:a8:1f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:2a:5c:81 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:e2:51:ba Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:b1:28:a1 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:ba:31:22:57:8a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:de:b6:ee:ef:43:4f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.128609 4979 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.129006 4979 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.130426 4979 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.130799 4979 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.130852 4979 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.131263 4979 topology_manager.go:138] "Creating topology manager with none policy" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.131285 4979 container_manager_linux.go:303] "Creating device plugin manager" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.131566 4979 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.131605 4979 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.131926 4979 state_mem.go:36] "Initialized new in-memory state store" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.132107 4979 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.133262 4979 kubelet.go:418] "Attempting to sync node with API server" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.133324 4979 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.133352 4979 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.133376 4979 kubelet.go:324] "Adding apiserver pod source" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.133396 4979 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.134770 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.134869 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.134893 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.134966 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.135954 4979 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.136343 4979 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137438 4979 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137941 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137964 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137972 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137979 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137990 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.137998 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138005 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138016 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138025 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138032 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138043 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138049 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138431 4979 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.138883 4979 server.go:1280] "Started kubelet" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.139088 4979 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.139146 4979 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.139897 4979 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.140102 4979 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 04 11:42:56 crc systemd[1]: Started Kubernetes Kubelet. Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.141706 4979 server.go:460] "Adding debug handlers to kubelet server" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.141897 4979 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.141932 4979 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.142484 4979 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.142507 4979 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.142476 4979 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 06:03:21.809899945 +0000 UTC Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.142520 4979 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.142575 4979 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.143887 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="200ms" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.143914 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.144040 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.145160 4979 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.74:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e006def1c43a9 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 11:42:56.138855337 +0000 UTC m=+0.413151131,LastTimestamp:2025-12-04 11:42:56.138855337 +0000 UTC m=+0.413151131,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.148263 4979 factory.go:55] Registering systemd factory Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.148330 4979 factory.go:221] Registration of the systemd container factory successfully Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.151614 4979 factory.go:153] Registering CRI-O factory Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.151686 4979 factory.go:221] Registration of the crio container factory successfully Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.151795 4979 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.153191 4979 factory.go:103] Registering Raw factory Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.153380 4979 manager.go:1196] Started watching for new ooms in manager Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.156494 4979 manager.go:319] Starting recovery of all containers Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157117 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157209 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157229 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157246 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157262 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157276 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157290 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157328 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157347 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157361 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157373 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157388 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157402 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157424 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157440 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157454 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157468 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157478 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157490 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157506 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157519 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157532 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157546 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157560 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157577 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157593 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157612 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157627 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157643 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157662 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157677 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157690 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157706 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157723 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157741 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157758 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157775 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157788 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157810 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157824 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157841 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157856 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157873 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157889 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157906 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157919 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157931 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157945 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157963 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157976 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.157991 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158007 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158026 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158072 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158091 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158106 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158120 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158134 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158173 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158186 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158198 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158216 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158252 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158266 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158277 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158290 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158317 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158332 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158344 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158357 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158373 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158390 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158401 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158414 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158428 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158438 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158452 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158467 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158481 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158491 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158503 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158516 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158532 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158547 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158564 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158576 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158591 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158603 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158619 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158633 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158644 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158661 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158675 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158688 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158701 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158713 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158727 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158739 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158751 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158762 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158776 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158790 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.158808 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159539 4979 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159569 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159595 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159612 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159628 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159642 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159660 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159678 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159698 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159714 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159731 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159751 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159765 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159779 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159797 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159815 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159831 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159846 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159864 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159884 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159901 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159916 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159930 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159944 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159958 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159973 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.159989 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160006 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160022 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160041 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160055 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160092 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160107 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160119 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160133 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160179 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160197 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160214 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160229 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160241 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160254 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160269 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160283 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160312 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160328 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160341 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160359 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160380 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160394 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160410 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160424 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160440 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160457 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160472 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160487 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160501 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160514 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160530 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160546 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160561 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160577 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160593 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160607 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160623 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160640 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160656 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160671 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160686 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160700 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160714 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160730 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160745 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160758 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160770 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160785 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160800 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160814 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160829 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160845 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160865 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160881 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160898 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160914 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160929 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160942 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160979 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.160994 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161008 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161025 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161038 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161052 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161065 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161078 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161093 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161106 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161119 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161132 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161145 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161157 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161168 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161180 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161193 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161207 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161220 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161231 4979 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161242 4979 reconstruct.go:97] "Volume reconstruction finished" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.161250 4979 reconciler.go:26] "Reconciler: start to sync state" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.182806 4979 manager.go:324] Recovery completed Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.195465 4979 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.196505 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.197683 4979 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.197729 4979 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.197756 4979 kubelet.go:2335] "Starting kubelet main sync loop" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.197879 4979 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.199715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.199768 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.199782 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.200778 4979 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.200798 4979 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.200820 4979 state_mem.go:36] "Initialized new in-memory state store" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.201111 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.201196 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.211771 4979 policy_none.go:49] "None policy: Start" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.212818 4979 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.212848 4979 state_mem.go:35] "Initializing new in-memory state store" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.243010 4979 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.270439 4979 manager.go:334] "Starting Device Plugin manager" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.270505 4979 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.270520 4979 server.go:79] "Starting device plugin registration server" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.271097 4979 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.271115 4979 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.271311 4979 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.271463 4979 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.271473 4979 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.281398 4979 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.298725 4979 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.298829 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.299857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.299893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.299908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300025 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300396 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300443 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300621 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300651 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300736 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300888 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.300920 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301145 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301160 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301170 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301440 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301560 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301766 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301806 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301865 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301891 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.301901 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302552 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302609 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302611 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302639 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302650 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302846 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302873 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.302906 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.303875 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.303908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.304026 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.304103 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.304107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.304335 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.304622 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.304670 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.305751 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.305798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.305810 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.344712 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="400ms" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364465 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364538 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364561 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364588 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364616 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364641 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364658 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364678 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364698 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364722 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364738 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364753 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364768 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364783 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.364799 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.371560 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.373130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.373178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.373193 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.373228 4979 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.373811 4979 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.465910 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.465963 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.465979 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.465999 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466027 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466050 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466069 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466090 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466111 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466134 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466167 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466169 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466196 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466223 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466216 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466259 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466273 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466328 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466335 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466334 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466433 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466493 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466517 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466539 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466552 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466560 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466566 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466579 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466674 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.466697 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.574972 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.576595 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.576642 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.576652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.576678 4979 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.577323 4979 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.630727 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.652592 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.657261 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-7d693f110f57f503748ce57d078d39f31df6afb643c5d6b558a35b5e3115d49b WatchSource:0}: Error finding container 7d693f110f57f503748ce57d078d39f31df6afb643c5d6b558a35b5e3115d49b: Status 404 returned error can't find the container with id 7d693f110f57f503748ce57d078d39f31df6afb643c5d6b558a35b5e3115d49b Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.658550 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.671024 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5fbf98a028c4866291c485c9a77d7c7a02c73c3856eb16546506f5837d6bfe38 WatchSource:0}: Error finding container 5fbf98a028c4866291c485c9a77d7c7a02c73c3856eb16546506f5837d6bfe38: Status 404 returned error can't find the container with id 5fbf98a028c4866291c485c9a77d7c7a02c73c3856eb16546506f5837d6bfe38 Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.674828 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-97bd5cf04d5eb9d8e4a6648a92bdbc961cfcee3a47b44f1d188f055d327fa502 WatchSource:0}: Error finding container 97bd5cf04d5eb9d8e4a6648a92bdbc961cfcee3a47b44f1d188f055d327fa502: Status 404 returned error can't find the container with id 97bd5cf04d5eb9d8e4a6648a92bdbc961cfcee3a47b44f1d188f055d327fa502 Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.682529 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.690682 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.710063 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a0aee4ab5f1f7c9bab475f469b8bd7576c6901d946195fa051ab43ea1316edae WatchSource:0}: Error finding container a0aee4ab5f1f7c9bab475f469b8bd7576c6901d946195fa051ab43ea1316edae: Status 404 returned error can't find the container with id a0aee4ab5f1f7c9bab475f469b8bd7576c6901d946195fa051ab43ea1316edae Dec 04 11:42:56 crc kubenswrapper[4979]: W1204 11:42:56.716056 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-ed284efdcc04bf8fd2791fee8a309feddcd1e6aae9d0922efbb240b5f6809595 WatchSource:0}: Error finding container ed284efdcc04bf8fd2791fee8a309feddcd1e6aae9d0922efbb240b5f6809595: Status 404 returned error can't find the container with id ed284efdcc04bf8fd2791fee8a309feddcd1e6aae9d0922efbb240b5f6809595 Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.745785 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="800ms" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.977860 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.979646 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.979701 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.979711 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:56 crc kubenswrapper[4979]: I1204 11:42:56.979744 4979 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 11:42:56 crc kubenswrapper[4979]: E1204 11:42:56.980397 4979 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.141244 4979 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.143584 4979 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-30 01:00:55.444602557 +0000 UTC Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.206158 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.206311 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a0aee4ab5f1f7c9bab475f469b8bd7576c6901d946195fa051ab43ea1316edae"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.208377 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713" exitCode=0 Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.208469 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.208505 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"97bd5cf04d5eb9d8e4a6648a92bdbc961cfcee3a47b44f1d188f055d327fa502"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.208649 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210211 4979 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2" exitCode=0 Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210274 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210327 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5fbf98a028c4866291c485c9a77d7c7a02c73c3856eb16546506f5837d6bfe38"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210481 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210810 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210849 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.210860 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.211701 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.211720 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.211730 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:57 crc kubenswrapper[4979]: W1204 11:42:57.212365 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:57 crc kubenswrapper[4979]: E1204 11:42:57.212464 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.212475 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.213618 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.213675 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.213687 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.215248 4979 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4" exitCode=0 Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.215333 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.215385 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7d693f110f57f503748ce57d078d39f31df6afb643c5d6b558a35b5e3115d49b"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.215486 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.216490 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.216513 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.216528 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.217656 4979 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5" exitCode=0 Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.217700 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.217736 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ed284efdcc04bf8fd2791fee8a309feddcd1e6aae9d0922efbb240b5f6809595"} Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.217891 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.218848 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.218873 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.218883 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:57 crc kubenswrapper[4979]: W1204 11:42:57.224838 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:57 crc kubenswrapper[4979]: E1204 11:42:57.224962 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:57 crc kubenswrapper[4979]: W1204 11:42:57.378033 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:57 crc kubenswrapper[4979]: E1204 11:42:57.378118 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:57 crc kubenswrapper[4979]: E1204 11:42:57.546941 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="1.6s" Dec 04 11:42:57 crc kubenswrapper[4979]: W1204 11:42:57.575332 4979 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.74:6443: connect: connection refused Dec 04 11:42:57 crc kubenswrapper[4979]: E1204 11:42:57.575437 4979 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.74:6443: connect: connection refused" logger="UnhandledError" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.780884 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.783064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.783171 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.783189 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:57 crc kubenswrapper[4979]: I1204 11:42:57.783371 4979 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 11:42:57 crc kubenswrapper[4979]: E1204 11:42:57.784997 4979 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.74:6443: connect: connection refused" node="crc" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.128655 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.144373 4979 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 10:45:56.366483245 +0000 UTC Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.144461 4979 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 887h2m58.222024993s for next certificate rotation Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.224647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.224709 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.224725 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.224906 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.225999 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.226055 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.226071 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.227847 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.227872 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.227883 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.227941 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.228805 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.228833 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.228844 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.231437 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.231497 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.231509 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.231522 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.236208 4979 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1" exitCode=0 Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.236337 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.236485 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.237648 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.237702 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.237717 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.239428 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c"} Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.239593 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.240782 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.240814 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:58 crc kubenswrapper[4979]: I1204 11:42:58.240824 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.244849 4979 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f" exitCode=0 Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.244983 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f"} Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.245344 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.246933 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.246988 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.247004 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.251594 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415"} Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.251715 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.251644 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.252815 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.252861 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.252874 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.252825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.252941 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.252951 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.386536 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.387780 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.387817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.387825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:42:59 crc kubenswrapper[4979]: I1204 11:42:59.387851 4979 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.197798 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.203384 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259662 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4"} Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259729 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259767 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259730 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e"} Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259849 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8"} Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259860 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca"} Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.259837 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.261370 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.261429 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.261441 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.261655 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.261715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.261730 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:00 crc kubenswrapper[4979]: I1204 11:43:00.837065 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.267318 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69"} Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.267465 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.267514 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.267554 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.267465 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.267644 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.272969 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273012 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273021 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273029 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273034 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273039 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273143 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273156 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.273166 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:01 crc kubenswrapper[4979]: I1204 11:43:01.752091 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.098729 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.270353 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.270429 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.270464 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.270527 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.270611 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.271988 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272019 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272076 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272088 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272051 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272039 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.272278 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:02 crc kubenswrapper[4979]: I1204 11:43:02.291362 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 04 11:43:03 crc kubenswrapper[4979]: I1204 11:43:03.273124 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:03 crc kubenswrapper[4979]: I1204 11:43:03.274707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:03 crc kubenswrapper[4979]: I1204 11:43:03.274760 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:03 crc kubenswrapper[4979]: I1204 11:43:03.274775 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.349025 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.349259 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.350674 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.350728 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.350743 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.529748 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.529978 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.531273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.531334 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:04 crc kubenswrapper[4979]: I1204 11:43:04.531347 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.099157 4979 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.099259 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.571505 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.571755 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.573268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.573370 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.573382 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:05 crc kubenswrapper[4979]: I1204 11:43:05.798661 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.153743 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.154043 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.155392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.155432 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.155444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:06 crc kubenswrapper[4979]: E1204 11:43:06.281742 4979 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.282968 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.284195 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.284234 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:06 crc kubenswrapper[4979]: I1204 11:43:06.284245 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:08 crc kubenswrapper[4979]: E1204 11:43:08.130904 4979 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 04 11:43:08 crc kubenswrapper[4979]: I1204 11:43:08.142419 4979 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 04 11:43:08 crc kubenswrapper[4979]: I1204 11:43:08.224954 4979 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 11:43:08 crc kubenswrapper[4979]: I1204 11:43:08.225033 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 11:43:08 crc kubenswrapper[4979]: I1204 11:43:08.230412 4979 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 04 11:43:08 crc kubenswrapper[4979]: I1204 11:43:08.230499 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.759177 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.759402 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.760089 4979 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.760165 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.768697 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.768721 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.768754 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:11 crc kubenswrapper[4979]: I1204 11:43:11.768837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.303573 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.304033 4979 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.304111 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.305176 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.305219 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.305228 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.351365 4979 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.351463 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.474322 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.487953 4979 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.517612 4979 csr.go:261] certificate signing request csr-pjgqv is approved, waiting to be issued Dec 04 11:43:12 crc kubenswrapper[4979]: I1204 11:43:12.525901 4979 csr.go:257] certificate signing request csr-pjgqv is issued Dec 04 11:43:13 crc kubenswrapper[4979]: E1204 11:43:13.225933 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.228570 4979 trace.go:236] Trace[450306614]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 11:42:58.881) (total time: 14346ms): Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[450306614]: ---"Objects listed" error: 14346ms (11:43:13.228) Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[450306614]: [14.346429668s] [14.346429668s] END Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.228628 4979 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.228580 4979 trace.go:236] Trace[529210193]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 11:42:58.884) (total time: 14344ms): Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[529210193]: ---"Objects listed" error: 14344ms (11:43:13.228) Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[529210193]: [14.344445084s] [14.344445084s] END Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.228684 4979 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.229498 4979 trace.go:236] Trace[364293732]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 11:42:59.361) (total time: 13868ms): Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[364293732]: ---"Objects listed" error: 13868ms (11:43:13.229) Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[364293732]: [13.8683142s] [13.8683142s] END Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.229536 4979 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 11:43:13 crc kubenswrapper[4979]: E1204 11:43:13.231004 4979 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.231622 4979 trace.go:236] Trace[1397273213]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (04-Dec-2025 11:42:59.281) (total time: 13949ms): Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[1397273213]: ---"Objects listed" error: 13949ms (11:43:13.231) Dec 04 11:43:13 crc kubenswrapper[4979]: Trace[1397273213]: [13.949987123s] [13.949987123s] END Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.231658 4979 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.269395 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.282822 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.286499 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.296487 4979 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 04 11:43:13 crc kubenswrapper[4979]: E1204 11:43:13.316030 4979 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.527125 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-04 11:38:12 +0000 UTC, rotation deadline is 2026-10-01 13:12:14.467348649 +0000 UTC Dec 04 11:43:13 crc kubenswrapper[4979]: I1204 11:43:13.527182 4979 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7225h29m0.940169462s for next certificate rotation Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.143242 4979 apiserver.go:52] "Watching apiserver" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.145984 4979 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.146455 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-smkn7","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.146983 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.147110 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.147287 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.147327 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.147366 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.147342 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.147404 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.147945 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.147964 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.148004 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.151101 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.151199 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.151361 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.151440 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.151774 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.151838 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.152946 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.153226 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.153398 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.153442 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.153629 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.159042 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.173389 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.187888 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.205705 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.214176 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.223882 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.235080 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.243136 4979 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.245987 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.256593 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.267115 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301744 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301798 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301818 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301837 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301856 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301878 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301895 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301912 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301930 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301949 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301964 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.301995 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302012 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302064 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302079 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302094 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302108 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302144 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302161 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302204 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302226 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302248 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302263 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302283 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302322 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302358 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302377 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302394 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302410 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302428 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302445 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302460 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302476 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302493 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302512 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302529 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302568 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302594 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302624 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302643 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302666 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302684 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302703 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302719 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302735 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302752 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302749 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302768 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302884 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302909 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302930 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302951 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302971 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.302994 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303012 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303034 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303052 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303058 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303157 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303183 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303203 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303224 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303250 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303271 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303289 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303331 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303352 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303369 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303387 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303404 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303422 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303421 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303440 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303450 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303461 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303502 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303528 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303548 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303570 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303587 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303603 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303651 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303673 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303694 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303712 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303732 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303749 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303749 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303767 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303766 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303783 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303799 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303816 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303832 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303852 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303869 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303886 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303902 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303919 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303937 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303954 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.304710 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303948 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.304997 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.303979 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305039 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305110 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305121 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305203 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305255 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305354 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305375 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305407 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305371 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305456 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305466 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305494 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305506 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305527 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305566 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305602 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305629 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305661 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305670 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305684 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305697 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305701 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305764 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305774 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305829 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305887 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305929 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305969 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305999 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306040 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306074 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306103 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306131 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306168 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306208 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306236 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306269 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306325 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306362 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306397 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306432 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306470 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306502 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306534 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306565 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306593 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306619 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306648 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306674 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306706 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306737 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306768 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306793 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306827 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306861 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307148 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307191 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307228 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307286 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307333 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307368 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307400 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307428 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307462 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307492 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307521 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307552 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307582 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307610 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307672 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307708 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307749 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307779 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307810 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307839 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307872 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307903 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307931 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307960 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307994 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308052 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308080 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308108 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308136 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308168 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308199 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308231 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308255 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308311 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.309414 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.310393 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.310448 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.310481 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.310506 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.310569 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305774 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.313075 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.305800 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.306407 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307013 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307155 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307370 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307758 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307829 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.307994 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308157 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.308762 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.309190 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.311419 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.311474 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.312140 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.312316 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.312755 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.312800 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.313231 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:43:14.81319047 +0000 UTC m=+19.087486274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.313343 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.309028 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.313369 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.313255 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.313447 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.313490 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314154 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314327 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314368 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314407 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314707 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314724 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.314751 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.315081 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.315129 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.312198 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.315538 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.315806 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.315822 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.315880 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316220 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316321 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316365 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316532 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316266 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316783 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316823 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.316925 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317042 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317292 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317501 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317473 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317579 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317839 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317991 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.317965 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318028 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318682 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318712 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318770 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318788 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319556 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319565 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318942 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318947 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318960 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319653 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.318970 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319734 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319841 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319887 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.320063 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.320247 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319071 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319100 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319127 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319149 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.320358 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.319199 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.320349 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.320607 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321012 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321011 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321571 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321639 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321649 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321740 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.321997 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322032 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322036 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322171 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322234 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322633 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322647 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.322959 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323019 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323193 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323225 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323262 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323270 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323291 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323453 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323464 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323515 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323517 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323589 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323649 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323676 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323747 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323745 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323775 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323693 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323836 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323866 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323893 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323920 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmq4z\" (UniqueName: \"kubernetes.io/projected/0dd48cc7-5561-4c17-9f53-f1e0c30129c2-kube-api-access-bmq4z\") pod \"node-resolver-smkn7\" (UID: \"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\") " pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323942 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323951 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323965 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.323989 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324012 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324033 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324001 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324041 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324104 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0dd48cc7-5561-4c17-9f53-f1e0c30129c2-hosts-file\") pod \"node-resolver-smkn7\" (UID: \"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\") " pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324142 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324173 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324204 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324234 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324360 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324395 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324429 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324525 4979 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324551 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.324563 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324546 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.324630 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:14.824608933 +0000 UTC m=+19.098904737 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324853 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.325329 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.325420 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:14.825393385 +0000 UTC m=+19.099689289 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325417 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325695 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.324568 4979 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325789 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325803 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325815 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325829 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325842 4979 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325855 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325867 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325877 4979 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325887 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325897 4979 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325912 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325927 4979 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325940 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325953 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325964 4979 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325974 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325985 4979 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.325997 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326007 4979 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326018 4979 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326028 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326038 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326048 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326058 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326068 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326078 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326087 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326097 4979 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326106 4979 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326115 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326126 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326138 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326149 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326159 4979 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326171 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326181 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326192 4979 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326332 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326354 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326370 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326383 4979 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326395 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326408 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326420 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326432 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326444 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326451 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326456 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326483 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326495 4979 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326610 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326623 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326632 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326642 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326652 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326663 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326674 4979 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326685 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326696 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326706 4979 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326717 4979 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326727 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326736 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326745 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326755 4979 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326765 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326783 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326794 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326806 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.326830 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.327352 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.327408 4979 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.327922 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.327952 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328060 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328078 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328778 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328794 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328810 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328380 4979 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328867 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328886 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328901 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328918 4979 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328933 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328978 4979 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.328992 4979 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329004 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329017 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329029 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329041 4979 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329053 4979 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329067 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329079 4979 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329094 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329107 4979 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329121 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329134 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329260 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329277 4979 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329331 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329436 4979 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329450 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329465 4979 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329478 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329494 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329510 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329524 4979 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329538 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329553 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329568 4979 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329581 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329595 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329608 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329622 4979 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329635 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329648 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329662 4979 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329675 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329689 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.329704 4979 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.330244 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.330207 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.330828 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.331364 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.331823 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.331970 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.332800 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.333345 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.333929 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.334786 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.335152 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.335562 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.336371 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.336415 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.336861 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.337269 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.338467 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.339661 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.339700 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.339719 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.339807 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:14.83977919 +0000 UTC m=+19.114075184 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.340889 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.340926 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.340948 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.341015 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:14.840991623 +0000 UTC m=+19.115287617 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.342027 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415" exitCode=255 Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.342158 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.342368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415"} Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.344067 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.345627 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.346019 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.347343 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.347664 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.348165 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.348444 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.349255 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.349546 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.349570 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.349721 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.349806 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350012 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350076 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350146 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350442 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350742 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350781 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.350795 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.351395 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.351955 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.352016 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.352317 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.352167 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.352530 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.352592 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.353512 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.353689 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.354428 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.354611 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.355087 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.355410 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.355446 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.355470 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.356351 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.358200 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.359196 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361073 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361209 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361240 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361272 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361497 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361549 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361538 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.361805 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362200 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362340 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362400 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362570 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362678 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362838 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.362900 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.363454 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.367003 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.368997 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.385650 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.393763 4979 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.406466 4979 scope.go:117] "RemoveContainer" containerID="1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.407946 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.408252 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.409624 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.412738 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.415791 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430640 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0dd48cc7-5561-4c17-9f53-f1e0c30129c2-hosts-file\") pod \"node-resolver-smkn7\" (UID: \"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\") " pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430793 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/0dd48cc7-5561-4c17-9f53-f1e0c30129c2-hosts-file\") pod \"node-resolver-smkn7\" (UID: \"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\") " pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430856 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmq4z\" (UniqueName: \"kubernetes.io/projected/0dd48cc7-5561-4c17-9f53-f1e0c30129c2-kube-api-access-bmq4z\") pod \"node-resolver-smkn7\" (UID: \"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\") " pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430879 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430900 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430942 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430958 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430974 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.430987 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431001 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431012 4979 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431024 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431037 4979 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431049 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431059 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431071 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431082 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431093 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431104 4979 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431117 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431129 4979 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431140 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431152 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431162 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431174 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431186 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431198 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431209 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431220 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431233 4979 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431244 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431257 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431269 4979 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431281 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431324 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431339 4979 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431351 4979 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431362 4979 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431373 4979 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431388 4979 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431402 4979 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431414 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431426 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431436 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431448 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431459 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431470 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431480 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431493 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431506 4979 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431517 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431528 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431539 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431553 4979 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431564 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431576 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431589 4979 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431601 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431633 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431644 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431657 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431667 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431678 4979 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431689 4979 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431699 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431709 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431722 4979 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431734 4979 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431745 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431756 4979 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431766 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431776 4979 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431786 4979 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431797 4979 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431757 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431813 4979 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431940 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431954 4979 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431982 4979 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.431990 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.432007 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.439853 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.462814 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmq4z\" (UniqueName: \"kubernetes.io/projected/0dd48cc7-5561-4c17-9f53-f1e0c30129c2-kube-api-access-bmq4z\") pod \"node-resolver-smkn7\" (UID: \"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\") " pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.466437 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.470817 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.475833 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 04 11:43:14 crc kubenswrapper[4979]: W1204 11:43:14.482560 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-6f243b0338eacbc177f2234805fdd27510c71393fe21f06cea4072b92e2c0f13 WatchSource:0}: Error finding container 6f243b0338eacbc177f2234805fdd27510c71393fe21f06cea4072b92e2c0f13: Status 404 returned error can't find the container with id 6f243b0338eacbc177f2234805fdd27510c71393fe21f06cea4072b92e2c0f13 Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.486684 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.492153 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-smkn7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.492618 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.504467 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: W1204 11:43:14.520967 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-ea9892a203138df12089147e2ed82ff60f67d5951edebc1c63ed858d151920f8 WatchSource:0}: Error finding container ea9892a203138df12089147e2ed82ff60f67d5951edebc1c63ed858d151920f8: Status 404 returned error can't find the container with id ea9892a203138df12089147e2ed82ff60f67d5951edebc1c63ed858d151920f8 Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.521266 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.573552 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-g4nk7"] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.573976 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.577947 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.577997 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.578046 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.578287 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.578495 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.591075 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.616795 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.630520 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.647912 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.660674 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.671198 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.684246 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.697994 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.709817 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.722261 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735001 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p5sn\" (UniqueName: \"kubernetes.io/projected/12339390-77cf-4906-80c3-ddb729e42132-kube-api-access-9p5sn\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735067 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-cni-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735095 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-kubelet\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735119 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-hostroot\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735144 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-os-release\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735160 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-socket-dir-parent\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735174 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-conf-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735226 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-etc-kubernetes\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735245 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-system-cni-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735258 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-k8s-cni-cncf-io\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735274 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-netns\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735289 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/12339390-77cf-4906-80c3-ddb729e42132-multus-daemon-config\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735365 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-cni-bin\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735382 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/12339390-77cf-4906-80c3-ddb729e42132-cni-binary-copy\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735398 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-cni-multus\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735447 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-cnibin\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.735461 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-multus-certs\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.836931 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837028 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-system-cni-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837052 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-k8s-cni-cncf-io\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837077 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-netns\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837096 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/12339390-77cf-4906-80c3-ddb729e42132-multus-daemon-config\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837113 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/12339390-77cf-4906-80c3-ddb729e42132-cni-binary-copy\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837131 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-cni-bin\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837172 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-cni-multus\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837172 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-k8s-cni-cncf-io\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837212 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-cni-multus\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.837216 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:43:15.837179858 +0000 UTC m=+20.111475662 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837245 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-netns\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.837286 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837319 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-system-cni-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.837351 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:15.837333532 +0000 UTC m=+20.111629336 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837343 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837446 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-cnibin\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.837469 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837493 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-multus-certs\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.837519 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:15.837511588 +0000 UTC m=+20.111807392 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837239 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-cni-bin\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837534 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-cnibin\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837470 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-run-multus-certs\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837582 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p5sn\" (UniqueName: \"kubernetes.io/projected/12339390-77cf-4906-80c3-ddb729e42132-kube-api-access-9p5sn\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837610 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-cni-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837647 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-kubelet\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837668 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-hostroot\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837697 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-os-release\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837716 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-socket-dir-parent\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837738 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-conf-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837759 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-etc-kubernetes\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837819 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-etc-kubernetes\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837851 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-hostroot\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.837877 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-os-release\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.838002 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/12339390-77cf-4906-80c3-ddb729e42132-multus-daemon-config\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.838046 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-conf-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.838071 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-socket-dir-parent\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.838076 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-host-var-lib-kubelet\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.838151 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/12339390-77cf-4906-80c3-ddb729e42132-multus-cni-dir\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.838205 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/12339390-77cf-4906-80c3-ddb729e42132-cni-binary-copy\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.856137 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p5sn\" (UniqueName: \"kubernetes.io/projected/12339390-77cf-4906-80c3-ddb729e42132-kube-api-access-9p5sn\") pod \"multus-g4nk7\" (UID: \"12339390-77cf-4906-80c3-ddb729e42132\") " pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.917166 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-g4nk7" Dec 04 11:43:14 crc kubenswrapper[4979]: W1204 11:43:14.927065 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod12339390_77cf_4906_80c3_ddb729e42132.slice/crio-86d41689075e2dca6c8a6ce8684d064696d12cdc66b107c4a0f444a7d5d27a15 WatchSource:0}: Error finding container 86d41689075e2dca6c8a6ce8684d064696d12cdc66b107c4a0f444a7d5d27a15: Status 404 returned error can't find the container with id 86d41689075e2dca6c8a6ce8684d064696d12cdc66b107c4a0f444a7d5d27a15 Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.938842 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.938899 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939067 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939091 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939105 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939170 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:15.939148878 +0000 UTC m=+20.213444682 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939456 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939491 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939504 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: E1204 11:43:14.939566 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:15.939545778 +0000 UTC m=+20.213841622 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.952924 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-95gj9"] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.953578 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.956580 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.956803 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.958584 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x6k66"] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.959516 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.961278 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.961823 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.961861 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.965417 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.965651 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.965663 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.966269 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-8twmz"] Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.966736 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.967385 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.971239 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.971480 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.971573 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.971782 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.971870 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.973464 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:14 crc kubenswrapper[4979]: I1204 11:43:14.982330 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.006609 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.017476 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.027823 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040159 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-ovn\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040207 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-script-lib\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040227 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-var-lib-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040386 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040478 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-cnibin\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040505 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c848f116-de32-47fa-9f47-5c0ff8204d29-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040528 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-etc-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040548 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-config\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040581 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-log-socket\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040612 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-ovn-kubernetes\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040635 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-netns\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040654 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-env-overrides\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040673 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovn-node-metrics-cert\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040718 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-tuning-conf-dir\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040739 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040758 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-bin\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040787 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-kubelet\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040806 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-os-release\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040822 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h6d6\" (UniqueName: \"kubernetes.io/projected/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-kube-api-access-6h6d6\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040848 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c848f116-de32-47fa-9f47-5c0ff8204d29-cni-binary-copy\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040868 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-systemd\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040902 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9x7gq\" (UniqueName: \"kubernetes.io/projected/c848f116-de32-47fa-9f47-5c0ff8204d29-kube-api-access-9x7gq\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040935 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-systemd-units\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040956 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-system-cni-dir\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040975 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-slash\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.040994 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-node-log\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.041014 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-netd\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.041577 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.058016 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.080077 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.100221 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.114132 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.129534 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142327 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9x7gq\" (UniqueName: \"kubernetes.io/projected/c848f116-de32-47fa-9f47-5c0ff8204d29-kube-api-access-9x7gq\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142667 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-systemd-units\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142756 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-node-log\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142830 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-netd\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142880 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-node-log\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142865 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-netd\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142904 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-system-cni-dir\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142999 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-slash\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143022 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-var-lib-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143040 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-ovn\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143060 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-script-lib\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143083 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143137 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9b48d97-5715-467e-b044-4812d6dfea63-proxy-tls\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143162 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-cnibin\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143182 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9b48d97-5715-467e-b044-4812d6dfea63-mcd-auth-proxy-config\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143203 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-config\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143223 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c848f116-de32-47fa-9f47-5c0ff8204d29-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143243 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-etc-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143270 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-log-socket\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143318 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-ovn-kubernetes\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-env-overrides\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143356 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovn-node-metrics-cert\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143376 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d9b48d97-5715-467e-b044-4812d6dfea63-rootfs\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143394 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x6hm\" (UniqueName: \"kubernetes.io/projected/d9b48d97-5715-467e-b044-4812d6dfea63-kube-api-access-8x6hm\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.142818 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-systemd-units\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143446 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-netns\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143421 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-netns\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143567 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143629 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-bin\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143653 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-tuning-conf-dir\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143681 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-kubelet\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143709 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-os-release\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143729 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6h6d6\" (UniqueName: \"kubernetes.io/projected/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-kube-api-access-6h6d6\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143745 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c848f116-de32-47fa-9f47-5c0ff8204d29-cni-binary-copy\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143763 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-systemd\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143787 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-etc-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143860 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143872 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-systemd\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.143920 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-cnibin\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144015 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-bin\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144063 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-kubelet\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144088 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144100 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-log-socket\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144146 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-tuning-conf-dir\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144167 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-os-release\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144182 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-ovn\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144189 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-ovn-kubernetes\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144204 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-var-lib-openvswitch\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144223 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-slash\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144274 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/c848f116-de32-47fa-9f47-5c0ff8204d29-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144437 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-config\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144574 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c848f116-de32-47fa-9f47-5c0ff8204d29-system-cni-dir\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144612 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-script-lib\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144715 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-env-overrides\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.144962 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c848f116-de32-47fa-9f47-5c0ff8204d29-cni-binary-copy\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.150607 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.150834 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovn-node-metrics-cert\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.164807 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.165124 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9x7gq\" (UniqueName: \"kubernetes.io/projected/c848f116-de32-47fa-9f47-5c0ff8204d29-kube-api-access-9x7gq\") pod \"multus-additional-cni-plugins-95gj9\" (UID: \"c848f116-de32-47fa-9f47-5c0ff8204d29\") " pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.165216 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6h6d6\" (UniqueName: \"kubernetes.io/projected/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-kube-api-access-6h6d6\") pod \"ovnkube-node-x6k66\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.186159 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.205415 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.224398 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.245438 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9b48d97-5715-467e-b044-4812d6dfea63-proxy-tls\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.246418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9b48d97-5715-467e-b044-4812d6dfea63-mcd-auth-proxy-config\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.246607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d9b48d97-5715-467e-b044-4812d6dfea63-rootfs\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.246210 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.246736 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d9b48d97-5715-467e-b044-4812d6dfea63-rootfs\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.246701 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x6hm\" (UniqueName: \"kubernetes.io/projected/d9b48d97-5715-467e-b044-4812d6dfea63-kube-api-access-8x6hm\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.247028 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d9b48d97-5715-467e-b044-4812d6dfea63-mcd-auth-proxy-config\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.249865 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d9b48d97-5715-467e-b044-4812d6dfea63-proxy-tls\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.267501 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x6hm\" (UniqueName: \"kubernetes.io/projected/d9b48d97-5715-467e-b044-4812d6dfea63-kube-api-access-8x6hm\") pod \"machine-config-daemon-8twmz\" (UID: \"d9b48d97-5715-467e-b044-4812d6dfea63\") " pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.271675 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.286958 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.301133 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.305055 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-95gj9" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.312535 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.319041 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.319097 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: W1204 11:43:15.333740 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf3c419d_ed2b_4fb7_bd28_5c6a5a8faa26.slice/crio-411c04f99f6258da1d397935c65aae70322c56cb7762d88705f89e71500e5dad WatchSource:0}: Error finding container 411c04f99f6258da1d397935c65aae70322c56cb7762d88705f89e71500e5dad: Status 404 returned error can't find the container with id 411c04f99f6258da1d397935c65aae70322c56cb7762d88705f89e71500e5dad Dec 04 11:43:15 crc kubenswrapper[4979]: W1204 11:43:15.351475 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9b48d97_5715_467e_b044_4812d6dfea63.slice/crio-f2ae75441020c1fa94889264363ce4fa49a260e3367cc912bde09eb3f53f4096 WatchSource:0}: Error finding container f2ae75441020c1fa94889264363ce4fa49a260e3367cc912bde09eb3f53f4096: Status 404 returned error can't find the container with id f2ae75441020c1fa94889264363ce4fa49a260e3367cc912bde09eb3f53f4096 Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.351672 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.371908 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerStarted","Data":"fec9e304d9dff170f3051005f16f8420f1f64c46dcd9be2223f72f6616bc23e4"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.385184 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.387511 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerStarted","Data":"dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.387566 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerStarted","Data":"86d41689075e2dca6c8a6ce8684d064696d12cdc66b107c4a0f444a7d5d27a15"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.406749 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.406826 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.406841 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"ea9892a203138df12089147e2ed82ff60f67d5951edebc1c63ed858d151920f8"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.426499 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-smkn7" event={"ID":"0dd48cc7-5561-4c17-9f53-f1e0c30129c2","Type":"ContainerStarted","Data":"02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.426556 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-smkn7" event={"ID":"0dd48cc7-5561-4c17-9f53-f1e0c30129c2","Type":"ContainerStarted","Data":"86468df5d71fc00cb215ef18294f028ad5d81bb860fc63c12c0d49567c908599"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.435701 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.435758 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"6f243b0338eacbc177f2234805fdd27510c71393fe21f06cea4072b92e2c0f13"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.437703 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"e3d0b4cbdefaffbc0747506aebc8ecd5e7176254424092ba1443e1b350688c4e"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.445036 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.447186 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.448018 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.453901 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"411c04f99f6258da1d397935c65aae70322c56cb7762d88705f89e71500e5dad"} Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.463958 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.529162 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.556637 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.578783 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.606185 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.622842 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.639027 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.655020 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.673980 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.692505 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.709292 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.722828 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.736589 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.752600 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:15Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.855243 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.855425 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.855482 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:43:17.855434419 +0000 UTC m=+22.129730223 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.855521 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.855639 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:17.855622734 +0000 UTC m=+22.129918538 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.855689 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.855856 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.855921 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:17.855911401 +0000 UTC m=+22.130207205 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.957402 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:15 crc kubenswrapper[4979]: I1204 11:43:15.957466 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957687 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957735 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957751 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957748 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957778 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957792 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957838 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:17.95781047 +0000 UTC m=+22.232106474 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:15 crc kubenswrapper[4979]: E1204 11:43:15.957867 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:17.957854651 +0000 UTC m=+22.232150675 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.066318 4979 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.188156 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.198256 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.198331 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.198256 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.198466 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.198636 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.198767 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.203609 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.204334 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.205443 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.206112 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.207467 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.208153 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.208825 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.209782 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.210077 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.210437 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.211629 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.212145 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.213236 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.213778 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.214262 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.217248 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.217829 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.218888 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.219446 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.220090 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.221978 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.222484 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.223609 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.224035 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.224444 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.225058 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.225580 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.226155 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.227227 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.228197 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.229279 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.230037 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.231154 4979 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.231271 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.233229 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.234630 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.235138 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.238216 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.238259 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.239632 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.240229 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.241512 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.242260 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.243680 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.244404 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.245534 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.247704 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.248262 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.249797 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.250483 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.252518 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.253166 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.253664 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.254056 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.255162 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.255835 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.256940 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.257529 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.258116 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.258225 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.267596 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.287424 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.302022 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.315732 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.329725 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.358782 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.375705 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.404872 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.419694 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.431154 4979 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.433368 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.433420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.433430 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.433742 4979 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.435479 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.443017 4979 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.443407 4979 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.444951 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.445016 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.445031 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.445058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.445076 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.457687 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" exitCode=0 Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.457830 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.459953 4979 generic.go:334] "Generic (PLEG): container finished" podID="c848f116-de32-47fa-9f47-5c0ff8204d29" containerID="4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c" exitCode=0 Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.460016 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerDied","Data":"4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.461642 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.461768 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.461884 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"f2ae75441020c1fa94889264363ce4fa49a260e3367cc912bde09eb3f53f4096"} Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.468380 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.468887 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.473160 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.473203 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.473222 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.473240 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.473253 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.479202 4979 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"etcd-crc\" already exists" pod="openshift-etcd/etcd-crc" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.497887 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.503347 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.507179 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.507286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.507378 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.507461 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.507582 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.515849 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.521312 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.525059 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.525190 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.525323 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.525409 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.525488 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.530417 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.568754 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.569123 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.589663 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.589718 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.589729 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.589747 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.589762 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.601048 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-x7qhm"] Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.601693 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.610784 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.610963 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.610806 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.611240 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.619765 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.636019 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: E1204 11:43:16.636185 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.642447 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.642491 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.642499 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.642516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.642528 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.666370 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.737054 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.744983 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.745025 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.745035 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.745055 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.745067 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.757737 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.770121 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c168d89e-7737-464e-a932-41993ad0c931-serviceca\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.770315 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c168d89e-7737-464e-a932-41993ad0c931-host\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.770370 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntxjs\" (UniqueName: \"kubernetes.io/projected/c168d89e-7737-464e-a932-41993ad0c931-kube-api-access-ntxjs\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.779067 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.807666 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.833407 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.847576 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.847626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.847636 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.847654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.847665 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.871452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c168d89e-7737-464e-a932-41993ad0c931-serviceca\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.871546 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c168d89e-7737-464e-a932-41993ad0c931-host\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.871608 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntxjs\" (UniqueName: \"kubernetes.io/projected/c168d89e-7737-464e-a932-41993ad0c931-kube-api-access-ntxjs\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.871709 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c168d89e-7737-464e-a932-41993ad0c931-host\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.872821 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/c168d89e-7737-464e-a932-41993ad0c931-serviceca\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.877680 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.905669 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntxjs\" (UniqueName: \"kubernetes.io/projected/c168d89e-7737-464e-a932-41993ad0c931-kube-api-access-ntxjs\") pod \"node-ca-x7qhm\" (UID: \"c168d89e-7737-464e-a932-41993ad0c931\") " pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.924202 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-x7qhm" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.936419 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:16 crc kubenswrapper[4979]: W1204 11:43:16.949815 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc168d89e_7737_464e_a932_41993ad0c931.slice/crio-e6d644c453a0b3e9e9cbe34435f4bde8a501049260088441bb54e9e467079c63 WatchSource:0}: Error finding container e6d644c453a0b3e9e9cbe34435f4bde8a501049260088441bb54e9e467079c63: Status 404 returned error can't find the container with id e6d644c453a0b3e9e9cbe34435f4bde8a501049260088441bb54e9e467079c63 Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.950891 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.950950 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.950970 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.950995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.951014 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:16Z","lastTransitionTime":"2025-12-04T11:43:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:16 crc kubenswrapper[4979]: I1204 11:43:16.975719 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.013888 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.055226 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.055290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.055347 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.055373 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.055387 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.063007 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.096752 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.135461 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.159419 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.159492 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.159538 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.159562 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.159578 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.179221 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.214642 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.254051 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.262174 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.262227 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.262238 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.262256 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.262269 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.295608 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.336070 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.365425 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.365467 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.365477 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.365492 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.365503 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.376045 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.414778 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.461172 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.467774 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.467826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.467842 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.467864 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.467879 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.472413 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.472463 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.472491 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.472504 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.472517 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.472529 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.474835 4979 generic.go:334] "Generic (PLEG): container finished" podID="c848f116-de32-47fa-9f47-5c0ff8204d29" containerID="00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5" exitCode=0 Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.474880 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerDied","Data":"00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.476908 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-x7qhm" event={"ID":"c168d89e-7737-464e-a932-41993ad0c931","Type":"ContainerStarted","Data":"14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.476947 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-x7qhm" event={"ID":"c168d89e-7737-464e-a932-41993ad0c931","Type":"ContainerStarted","Data":"e6d644c453a0b3e9e9cbe34435f4bde8a501049260088441bb54e9e467079c63"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.502253 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.537832 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.570436 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.570512 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.570525 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.570547 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.570569 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.576168 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.620999 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.655151 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.673417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.673464 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.673480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.673501 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.673517 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.702286 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.734530 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.776424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.776739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.776829 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.776924 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.777007 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.778240 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.815230 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.854351 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.880811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.880857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.880871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.880890 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.880905 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.881398 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.881590 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.881682 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.881686 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:43:21.881600182 +0000 UTC m=+26.155895986 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.881886 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.881943 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:21.881890559 +0000 UTC m=+26.156186483 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.882048 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.882136 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:21.882113165 +0000 UTC m=+26.156408979 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.896504 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.934017 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.976353 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.982522 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.982574 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982703 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982724 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982726 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982772 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982787 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982859 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:21.982836174 +0000 UTC m=+26.257131998 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982737 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:17 crc kubenswrapper[4979]: E1204 11:43:17.982907 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:21.982896565 +0000 UTC m=+26.257192379 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.983275 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.983329 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.983343 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.983362 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:17 crc kubenswrapper[4979]: I1204 11:43:17.983377 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:17Z","lastTransitionTime":"2025-12-04T11:43:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.016017 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.057173 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.085589 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.085629 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.085832 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.085847 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.085858 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.096990 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.188975 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.189039 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.189053 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.189073 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.189087 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.198667 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:18 crc kubenswrapper[4979]: E1204 11:43:18.198860 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.198703 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:18 crc kubenswrapper[4979]: E1204 11:43:18.198968 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.198897 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:18 crc kubenswrapper[4979]: E1204 11:43:18.199053 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.292536 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.292596 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.292609 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.292631 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.292644 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.395123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.395177 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.395195 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.395217 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.395231 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.487665 4979 generic.go:334] "Generic (PLEG): container finished" podID="c848f116-de32-47fa-9f47-5c0ff8204d29" containerID="75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80" exitCode=0 Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.487761 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerDied","Data":"75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.489454 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.498380 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.498433 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.498446 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.498467 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.498481 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.505822 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.523128 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.537594 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.550989 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.565728 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.579971 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.604570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.604659 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.604776 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.604888 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.604966 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.605987 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.638717 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.659022 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.674894 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.689364 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.706791 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.708699 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.708741 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.708753 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.708771 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.708785 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.719871 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.734209 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.749193 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.772699 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.795110 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.811608 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.811650 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.811660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.811674 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.811685 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.814246 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.858867 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.894562 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.914029 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.914072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.914081 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.914100 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.914111 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:18Z","lastTransitionTime":"2025-12-04T11:43:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.934218 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:18 crc kubenswrapper[4979]: I1204 11:43:18.973804 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.016734 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.017455 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.017506 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.017519 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.017537 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.017549 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.055500 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.096464 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.120645 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.120710 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.120727 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.120752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.120767 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.135046 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.175676 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.224274 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.224904 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.224917 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.224730 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.224935 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.225038 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.255146 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.293192 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.327040 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.327077 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.327086 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.327102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.327115 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.429534 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.429584 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.429597 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.429616 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.429632 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.499789 4979 generic.go:334] "Generic (PLEG): container finished" podID="c848f116-de32-47fa-9f47-5c0ff8204d29" containerID="20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f" exitCode=0 Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.499871 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerDied","Data":"20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.507116 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.520703 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.533953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.533983 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.533994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.534010 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.534021 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.534383 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.556161 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.576156 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.590661 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.605252 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.621554 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.637007 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.637081 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.637092 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.637142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.637157 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.638050 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.654961 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.695847 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.735875 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.739673 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.739705 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.739715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.739730 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.739741 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.774240 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.813589 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.842769 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.842855 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.842870 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.842894 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.842908 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.854598 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.895316 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.946535 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.946587 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.946597 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.946612 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:19 crc kubenswrapper[4979]: I1204 11:43:19.946622 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:19Z","lastTransitionTime":"2025-12-04T11:43:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.050013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.050070 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.050082 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.050113 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.050125 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.152839 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.152884 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.152893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.152910 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.152923 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.198589 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.198667 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.198605 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:20 crc kubenswrapper[4979]: E1204 11:43:20.198774 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:20 crc kubenswrapper[4979]: E1204 11:43:20.198851 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:20 crc kubenswrapper[4979]: E1204 11:43:20.198926 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.256820 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.257374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.257385 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.257415 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.257428 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.360571 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.361513 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.361560 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.361585 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.361599 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.464064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.464105 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.464114 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.464130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.464140 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.514973 4979 generic.go:334] "Generic (PLEG): container finished" podID="c848f116-de32-47fa-9f47-5c0ff8204d29" containerID="78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424" exitCode=0 Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.515029 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerDied","Data":"78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.532133 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.548397 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.566346 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.566391 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.566400 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.566417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.566428 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.568933 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.584439 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.599782 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.612970 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.625841 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.644964 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.660935 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.669132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.669250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.669264 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.669284 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.669376 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.673338 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.695111 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.705693 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.727598 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.749938 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.762325 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:20Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.772128 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.772178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.772189 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.772212 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.772226 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.875006 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.875041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.875050 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.875064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.875074 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.979991 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.980040 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.980052 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.980075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:20 crc kubenswrapper[4979]: I1204 11:43:20.980089 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:20Z","lastTransitionTime":"2025-12-04T11:43:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.083250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.083334 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.083348 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.083369 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.083381 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.186057 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.186105 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.186116 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.186134 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.186152 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.288649 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.288706 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.288718 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.288738 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.288750 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.391968 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.392019 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.392028 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.392046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.392060 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.495542 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.495615 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.495637 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.495667 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.495689 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.521028 4979 generic.go:334] "Generic (PLEG): container finished" podID="c848f116-de32-47fa-9f47-5c0ff8204d29" containerID="77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd" exitCode=0 Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.521073 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerDied","Data":"77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.539529 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.557757 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.580031 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.599098 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.611616 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.611673 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.611683 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.611705 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.611717 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.647544 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.671200 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.686969 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.710237 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.714660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.714698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.714710 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.714726 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.714738 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.731336 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.746947 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.760022 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.775137 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.793220 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.812201 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.817801 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.817836 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.817852 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.817868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.817880 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.827126 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:21Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.920180 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.920230 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.920239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.920256 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.920266 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:21Z","lastTransitionTime":"2025-12-04T11:43:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.927635 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:21 crc kubenswrapper[4979]: E1204 11:43:21.927788 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:43:29.927756982 +0000 UTC m=+34.202052846 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.927841 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:21 crc kubenswrapper[4979]: I1204 11:43:21.927937 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:21 crc kubenswrapper[4979]: E1204 11:43:21.927968 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:21 crc kubenswrapper[4979]: E1204 11:43:21.928031 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:29.928012599 +0000 UTC m=+34.202308403 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:21 crc kubenswrapper[4979]: E1204 11:43:21.928165 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:21 crc kubenswrapper[4979]: E1204 11:43:21.928346 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:29.928274335 +0000 UTC m=+34.202570329 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.023438 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.023482 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.023492 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.023510 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.023523 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.029115 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.029223 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029370 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029401 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029370 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029421 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029430 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029445 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029500 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:30.029473727 +0000 UTC m=+34.303769601 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.029518 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:30.029510448 +0000 UTC m=+34.303806332 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.126204 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.126248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.126259 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.126274 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.126287 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.198811 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.198892 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.199276 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.198988 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.199608 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:22 crc kubenswrapper[4979]: E1204 11:43:22.200083 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.229669 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.229721 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.229740 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.229765 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.229784 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.332605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.332653 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.332662 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.332678 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.332690 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.434700 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.434757 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.434767 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.434784 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.434795 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.530174 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" event={"ID":"c848f116-de32-47fa-9f47-5c0ff8204d29","Type":"ContainerStarted","Data":"13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.535816 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.536205 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.536502 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.536543 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.536555 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.536576 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.536593 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.558900 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.562009 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.583187 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.597583 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.614382 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.629604 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.639072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.639114 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.639123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.639137 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.639149 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.645976 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.659837 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.674139 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.703227 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.720265 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.734737 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.742054 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.742102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.742116 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.742145 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.742165 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.746805 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.764505 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.817070 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.830049 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.843495 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.844475 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.844517 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.844528 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.844545 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.844557 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.856317 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.870991 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.884253 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.897643 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.906521 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.924943 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.944725 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.948286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.948345 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.948359 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.948375 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.948387 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:22Z","lastTransitionTime":"2025-12-04T11:43:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.957412 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.971394 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:22 crc kubenswrapper[4979]: I1204 11:43:22.986590 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:22Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.006394 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.023536 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.038169 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.051021 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.051813 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.051911 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.051935 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.051951 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.051962 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.155255 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.155386 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.155413 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.155442 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.155468 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.258714 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.258769 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.258781 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.258800 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.258812 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.362038 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.362102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.362116 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.362143 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.362156 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.465178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.465234 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.465245 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.465267 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.465279 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.539930 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.540612 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.566694 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.568027 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.568066 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.568082 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.568103 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.568116 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.584207 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.602864 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.619419 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.637317 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.654580 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.671127 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.671184 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.671201 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.671253 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.671279 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.680876 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.715893 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.751056 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.774254 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.774320 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.774333 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.774352 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.774366 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.781126 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.799894 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.820327 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.836498 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.853786 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.865902 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.877063 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.877146 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.877165 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.877184 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.877195 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.881218 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:23Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.980097 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.980144 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.980154 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.980171 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:23 crc kubenswrapper[4979]: I1204 11:43:23.980181 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:23Z","lastTransitionTime":"2025-12-04T11:43:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.083722 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.083757 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.083767 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.083786 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.083798 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.189502 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.189550 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.189562 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.189584 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.189597 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.198173 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:24 crc kubenswrapper[4979]: E1204 11:43:24.198332 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.198432 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:24 crc kubenswrapper[4979]: E1204 11:43:24.198506 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.198611 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:24 crc kubenswrapper[4979]: E1204 11:43:24.198732 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.292857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.292908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.292918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.292935 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.292948 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.354795 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.374711 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.386806 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.395552 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.395601 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.395614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.395635 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.395651 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.398868 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.426674 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.440782 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.453784 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.467814 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.482578 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.496428 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.498064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.498097 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.498107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.498124 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.498136 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.506370 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.517900 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.527622 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.545717 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/0.log" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.546025 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.548575 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff" exitCode=1 Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.548626 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.549285 4979 scope.go:117] "RemoveContainer" containerID="3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.572384 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.589708 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.601155 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.601221 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.601239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.601265 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.601277 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.602745 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.619683 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.632985 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.649149 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.666459 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.680244 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.704604 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.704659 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.704673 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.704692 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.704703 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.704850 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.726276 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"message\\\":\\\"go:160\\\\nI1204 11:43:24.320087 6289 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320162 6289 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320320 6289 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 11:43:24.320348 6289 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 11:43:24.320400 6289 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 11:43:24.320448 6289 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 11:43:24.320465 6289 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 11:43:24.320463 6289 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 11:43:24.320481 6289 factory.go:656] Stopping watch factory\\\\nI1204 11:43:24.320484 6289 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 11:43:24.320499 6289 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 11:43:24.320504 6289 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.740156 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.753413 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.770812 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.787910 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.805381 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.807063 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.807099 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.807111 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.807130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.807144 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.822939 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.838592 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:24Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.910353 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.910437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.910456 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.910520 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:24 crc kubenswrapper[4979]: I1204 11:43:24.910543 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:24Z","lastTransitionTime":"2025-12-04T11:43:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.013644 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.013709 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.013726 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.013747 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.013766 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.116596 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.116649 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.116665 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.116684 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.116697 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.221043 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.221148 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.221183 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.221208 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.221221 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.325460 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.325500 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.325511 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.325529 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.325540 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.428851 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.428913 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.428927 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.428949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.428970 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.531742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.531800 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.531813 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.531833 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.531854 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.553278 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/0.log" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.555395 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.555520 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.575862 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.588755 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.613358 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.634908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.634982 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.634996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.635019 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.635034 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.637536 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"message\\\":\\\"go:160\\\\nI1204 11:43:24.320087 6289 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320162 6289 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320320 6289 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 11:43:24.320348 6289 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 11:43:24.320400 6289 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 11:43:24.320448 6289 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 11:43:24.320465 6289 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 11:43:24.320463 6289 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 11:43:24.320481 6289 factory.go:656] Stopping watch factory\\\\nI1204 11:43:24.320484 6289 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 11:43:24.320499 6289 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 11:43:24.320504 6289 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.652269 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.670247 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.687246 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.700051 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.713432 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.730449 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.737273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.737659 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.737749 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.737846 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.737932 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.747877 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.761568 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.780973 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.793955 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.810887 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:25Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.840597 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.840975 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.841064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.841150 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.841230 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.944085 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.944404 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.944470 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.944536 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:25 crc kubenswrapper[4979]: I1204 11:43:25.944592 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:25Z","lastTransitionTime":"2025-12-04T11:43:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.048000 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.048046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.048056 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.048075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.048093 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.150516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.150623 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.150637 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.150659 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.150673 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.198639 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.198782 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.198898 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.198894 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.198934 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.199356 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.225778 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.247084 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"message\\\":\\\"go:160\\\\nI1204 11:43:24.320087 6289 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320162 6289 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320320 6289 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 11:43:24.320348 6289 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 11:43:24.320400 6289 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 11:43:24.320448 6289 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 11:43:24.320465 6289 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 11:43:24.320463 6289 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 11:43:24.320481 6289 factory.go:656] Stopping watch factory\\\\nI1204 11:43:24.320484 6289 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 11:43:24.320499 6289 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 11:43:24.320504 6289 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.253117 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.253184 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.253197 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.253215 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.253225 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.261441 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.276648 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.292131 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.307560 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.322467 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.338486 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.357496 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.357558 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.357570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.357595 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.357610 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.361033 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.375374 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.389832 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.401247 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.415887 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.429142 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.440677 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.462031 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.462083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.462096 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.462116 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.462133 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.562553 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/1.log" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.566258 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/0.log" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.568083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.568279 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.568645 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.568709 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.568735 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.573236 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177" exitCode=1 Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.573284 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.573366 4979 scope.go:117] "RemoveContainer" containerID="3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.574139 4979 scope.go:117] "RemoveContainer" containerID="e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.574488 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.599380 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.634989 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"message\\\":\\\"go:160\\\\nI1204 11:43:24.320087 6289 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320162 6289 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320320 6289 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 11:43:24.320348 6289 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 11:43:24.320400 6289 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 11:43:24.320448 6289 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 11:43:24.320465 6289 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 11:43:24.320463 6289 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 11:43:24.320481 6289 factory.go:656] Stopping watch factory\\\\nI1204 11:43:24.320484 6289 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 11:43:24.320499 6289 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 11:43:24.320504 6289 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.656709 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.671918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.672156 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.672167 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.672187 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.672201 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.680436 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.697932 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.719912 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.736081 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.751944 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.766600 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.774731 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.774785 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.774804 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.774825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.774842 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.783365 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.797813 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.813028 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.813871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.813912 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.813929 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.813950 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.813963 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.827361 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.831281 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.832431 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.832491 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.832504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.832527 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.832541 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.850392 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.851823 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.855987 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.856025 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.856037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.856054 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.856064 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.865732 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.870078 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.874694 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.874750 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.874761 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.874778 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.874790 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.889128 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.893974 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.894033 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.894053 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.894075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.894087 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.910348 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:26 crc kubenswrapper[4979]: E1204 11:43:26.910534 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.912698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.912735 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.912751 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.912771 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:26 crc kubenswrapper[4979]: I1204 11:43:26.912784 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:26Z","lastTransitionTime":"2025-12-04T11:43:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.015831 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.015919 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.015936 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.015961 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.015981 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.118782 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.118839 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.118889 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.118912 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.118925 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.221892 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.221940 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.221950 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.221964 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.221973 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.324736 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.324789 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.324801 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.324822 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.324836 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.427947 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.428055 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.428075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.428100 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.428117 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.440494 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c"] Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.441030 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.443441 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.443656 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.460200 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.480543 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.494004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f575g\" (UniqueName: \"kubernetes.io/projected/a19f9554-b031-4d5c-a4aa-45f519d09945-kube-api-access-f575g\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.494111 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a19f9554-b031-4d5c-a4aa-45f519d09945-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.494269 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a19f9554-b031-4d5c-a4aa-45f519d09945-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.494549 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a19f9554-b031-4d5c-a4aa-45f519d09945-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.495793 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.511921 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.525385 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.531264 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.531286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.531317 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.531331 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.531342 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.542688 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.556338 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.569880 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.578069 4979 scope.go:117] "RemoveContainer" containerID="e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177" Dec 04 11:43:27 crc kubenswrapper[4979]: E1204 11:43:27.578247 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.583001 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.594400 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.595712 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a19f9554-b031-4d5c-a4aa-45f519d09945-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.595765 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a19f9554-b031-4d5c-a4aa-45f519d09945-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.595797 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f575g\" (UniqueName: \"kubernetes.io/projected/a19f9554-b031-4d5c-a4aa-45f519d09945-kube-api-access-f575g\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.595815 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a19f9554-b031-4d5c-a4aa-45f519d09945-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.596602 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a19f9554-b031-4d5c-a4aa-45f519d09945-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.597382 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a19f9554-b031-4d5c-a4aa-45f519d09945-env-overrides\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.603667 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a19f9554-b031-4d5c-a4aa-45f519d09945-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.610232 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.619256 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f575g\" (UniqueName: \"kubernetes.io/projected/a19f9554-b031-4d5c-a4aa-45f519d09945-kube-api-access-f575g\") pod \"ovnkube-control-plane-749d76644c-jf25c\" (UID: \"a19f9554-b031-4d5c-a4aa-45f519d09945\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.624596 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.634851 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.634922 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.634931 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.634949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.634963 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.651406 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.673461 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3aa62ddf868bffa23465456a09529954a2c4d1c5fa3c613077c0a905eb580bff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"message\\\":\\\"go:160\\\\nI1204 11:43:24.320087 6289 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320162 6289 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1204 11:43:24.320320 6289 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1204 11:43:24.320348 6289 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1204 11:43:24.320400 6289 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1204 11:43:24.320448 6289 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1204 11:43:24.320465 6289 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1204 11:43:24.320463 6289 handler.go:208] Removed *v1.Node event handler 2\\\\nI1204 11:43:24.320481 6289 factory.go:656] Stopping watch factory\\\\nI1204 11:43:24.320484 6289 handler.go:208] Removed *v1.Node event handler 7\\\\nI1204 11:43:24.320499 6289 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1204 11:43:24.320504 6289 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.685856 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.701376 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.723576 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.738476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.738538 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.738554 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.738576 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.738590 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.739625 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.754226 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.757039 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" Dec 04 11:43:27 crc kubenswrapper[4979]: W1204 11:43:27.772532 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda19f9554_b031_4d5c_a4aa_45f519d09945.slice/crio-8f8b99cabb91e6896cff160e34f0accadbe4c36fcbf36da092c9200f7c67f9ae WatchSource:0}: Error finding container 8f8b99cabb91e6896cff160e34f0accadbe4c36fcbf36da092c9200f7c67f9ae: Status 404 returned error can't find the container with id 8f8b99cabb91e6896cff160e34f0accadbe4c36fcbf36da092c9200f7c67f9ae Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.773993 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.793257 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.817526 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.833802 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.842497 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.842591 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.842614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.842638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.842655 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.847612 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.863407 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.877940 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.891387 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.904820 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.915680 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.931319 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.945005 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.945720 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.945758 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.945771 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.945792 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.945804 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:27Z","lastTransitionTime":"2025-12-04T11:43:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:27 crc kubenswrapper[4979]: I1204 11:43:27.959543 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:27Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.047949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.047985 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.047995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.048009 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.048021 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.152227 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.152321 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.152339 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.152364 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.152384 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.198160 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.198230 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:28 crc kubenswrapper[4979]: E1204 11:43:28.198583 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:28 crc kubenswrapper[4979]: E1204 11:43:28.198809 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.201567 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:28 crc kubenswrapper[4979]: E1204 11:43:28.201760 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.255861 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.255918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.255931 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.255952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.255965 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.359448 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.359502 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.359511 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.359525 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.359536 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.462570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.462626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.462636 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.462652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.462663 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.565394 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.565447 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.565457 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.565480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.565496 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.581143 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-td9cq"] Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.581804 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:28 crc kubenswrapper[4979]: E1204 11:43:28.581889 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.587622 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" event={"ID":"a19f9554-b031-4d5c-a4aa-45f519d09945","Type":"ContainerStarted","Data":"a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.587670 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" event={"ID":"a19f9554-b031-4d5c-a4aa-45f519d09945","Type":"ContainerStarted","Data":"8f8b99cabb91e6896cff160e34f0accadbe4c36fcbf36da092c9200f7c67f9ae"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.589884 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/1.log" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.598208 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.613348 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.624845 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.639255 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.654101 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.668928 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.668985 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.668995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.669013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.669024 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.669271 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.680397 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.693347 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.704782 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.713248 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.713468 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsp9b\" (UniqueName: \"kubernetes.io/projected/e031b763-d441-46af-a2af-ad62bd130a6c-kube-api-access-hsp9b\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.731558 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.754529 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.772194 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.773101 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.773138 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.773147 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.773165 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.773177 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.791542 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.808025 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.814737 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsp9b\" (UniqueName: \"kubernetes.io/projected/e031b763-d441-46af-a2af-ad62bd130a6c-kube-api-access-hsp9b\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.815539 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:28 crc kubenswrapper[4979]: E1204 11:43:28.815636 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:28 crc kubenswrapper[4979]: E1204 11:43:28.815725 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:43:29.315705684 +0000 UTC m=+33.590001488 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.828585 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.841696 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsp9b\" (UniqueName: \"kubernetes.io/projected/e031b763-d441-46af-a2af-ad62bd130a6c-kube-api-access-hsp9b\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.848638 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.863952 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.876910 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.876955 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.876963 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.876979 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.876990 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.979408 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.979466 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.979481 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.979504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:28 crc kubenswrapper[4979]: I1204 11:43:28.979520 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:28Z","lastTransitionTime":"2025-12-04T11:43:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.082186 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.082230 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.082242 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.082262 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.082274 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.185590 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.185627 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.185636 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.185649 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.185660 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.287981 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.288026 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.288037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.288054 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.288066 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.320553 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:29 crc kubenswrapper[4979]: E1204 11:43:29.320706 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:29 crc kubenswrapper[4979]: E1204 11:43:29.320773 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:43:30.320751958 +0000 UTC m=+34.595047762 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.390959 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.391026 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.391038 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.391053 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.391063 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.493583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.493624 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.493635 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.493652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.493667 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.596755 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.596818 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.596837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.596878 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.596896 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.600543 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" event={"ID":"a19f9554-b031-4d5c-a4aa-45f519d09945","Type":"ContainerStarted","Data":"5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.615783 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.633169 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.651124 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.666713 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.680278 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.695141 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.698972 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.699023 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.699037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.699052 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.699061 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.708919 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.731617 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.752028 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.764442 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.778079 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.794002 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.802744 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.802788 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.802800 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.802820 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.802832 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.814441 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.834364 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.847238 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.860377 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.875802 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:29Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.906119 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.906153 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.906163 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.906179 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:29 crc kubenswrapper[4979]: I1204 11:43:29.906189 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:29Z","lastTransitionTime":"2025-12-04T11:43:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.009281 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.009341 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.009352 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.009368 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.009379 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.027747 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.027902 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.027941 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:43:46.027913468 +0000 UTC m=+50.302209272 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.027992 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.028014 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.028037 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:46.028027771 +0000 UTC m=+50.302323575 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.028108 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.028139 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:46.028131223 +0000 UTC m=+50.302427027 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.112586 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.112630 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.112641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.112657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.112669 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.129452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.129528 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129662 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129681 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129694 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129699 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129751 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129769 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129753 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:46.129736865 +0000 UTC m=+50.404032669 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.129874 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:43:46.129850809 +0000 UTC m=+50.404146633 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.198627 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.198744 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.198805 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.198845 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.199072 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.199183 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.199218 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.199274 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.215929 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.215996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.216016 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.216041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.216060 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.318964 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.319038 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.319055 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.319074 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.319085 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.331799 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.332025 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: E1204 11:43:30.332131 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:43:32.332105688 +0000 UTC m=+36.606401492 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.422733 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.422839 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.422863 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.422889 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.422912 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.526318 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.526405 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.526416 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.526436 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.526452 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.629646 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.629716 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.629734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.629763 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.629780 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.732797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.732846 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.732856 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.732879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.732901 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.835004 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.835070 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.835087 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.835140 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.835156 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.938599 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.938660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.938675 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.938699 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:30 crc kubenswrapper[4979]: I1204 11:43:30.938711 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:30Z","lastTransitionTime":"2025-12-04T11:43:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.041374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.041430 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.041444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.041466 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.041479 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.143561 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.143610 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.143621 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.143640 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.143655 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.246420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.246480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.246493 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.246513 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.246529 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.349519 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.349588 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.349601 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.349619 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.349629 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.452899 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.452966 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.452984 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.453007 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.453021 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.556128 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.556217 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.556239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.556278 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.556341 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.556451 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.557287 4979 scope.go:117] "RemoveContainer" containerID="e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177" Dec 04 11:43:31 crc kubenswrapper[4979]: E1204 11:43:31.557495 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.660138 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.660202 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.660218 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.660243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.660262 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.763395 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.763449 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.763464 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.763485 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.763499 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.866868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.866941 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.866953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.866971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.866983 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.970206 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.970273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.970287 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.970322 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:31 crc kubenswrapper[4979]: I1204 11:43:31.970335 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:31Z","lastTransitionTime":"2025-12-04T11:43:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.072893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.072952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.072968 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.072989 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.072999 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.175213 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.175261 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.175272 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.175290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.175332 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.198279 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:32 crc kubenswrapper[4979]: E1204 11:43:32.198427 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.198621 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:32 crc kubenswrapper[4979]: E1204 11:43:32.199403 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.199464 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.199505 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:32 crc kubenswrapper[4979]: E1204 11:43:32.199567 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:32 crc kubenswrapper[4979]: E1204 11:43:32.199629 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.278200 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.278240 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.278254 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.278271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.278286 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.354560 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:32 crc kubenswrapper[4979]: E1204 11:43:32.354885 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:32 crc kubenswrapper[4979]: E1204 11:43:32.354984 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:43:36.354954432 +0000 UTC m=+40.629250266 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.381079 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.381113 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.381121 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.381137 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.381147 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.483575 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.483621 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.483632 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.483651 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.483664 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.587282 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.587359 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.587373 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.587392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.587406 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.690222 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.690277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.690286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.690324 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.690335 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.793102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.793160 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.793176 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.793197 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.793211 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.896461 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.896529 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.896542 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.896564 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:32 crc kubenswrapper[4979]: I1204 11:43:32.896578 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:32Z","lastTransitionTime":"2025-12-04T11:43:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.000468 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.000540 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.000550 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.000578 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.000591 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.103865 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.103917 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.103928 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.103947 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.103961 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.206957 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.207020 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.207042 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.207066 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.207085 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.310752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.310826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.310837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.310857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.310870 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.414083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.414137 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.414147 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.414164 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.414183 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.516656 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.516704 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.516715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.516735 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.516745 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.618544 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.618591 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.618603 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.618625 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.618636 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.722111 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.722172 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.722187 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.722208 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.722221 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.825512 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.825565 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.825581 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.825604 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.825620 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.928952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.929027 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.929046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.929070 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:33 crc kubenswrapper[4979]: I1204 11:43:33.929091 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:33Z","lastTransitionTime":"2025-12-04T11:43:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.032600 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.032676 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.032686 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.032706 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.032719 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.135639 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.135690 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.135699 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.135715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.135724 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.198544 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.198545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:34 crc kubenswrapper[4979]: E1204 11:43:34.198718 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.198701 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.198908 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:34 crc kubenswrapper[4979]: E1204 11:43:34.199531 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:34 crc kubenswrapper[4979]: E1204 11:43:34.199606 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:34 crc kubenswrapper[4979]: E1204 11:43:34.199663 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.238906 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.238972 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.238981 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.238997 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.239007 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.342251 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.342293 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.342347 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.342363 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.342422 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.444929 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.444975 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.444986 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.445000 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.445009 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.548695 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.549130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.549205 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.549345 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.549458 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.652126 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.652198 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.652210 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.652226 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.652241 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.755788 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.755869 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.755895 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.755929 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.755959 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.858873 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.858920 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.858937 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.858957 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.858971 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.961852 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.961918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.961931 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.961952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:34 crc kubenswrapper[4979]: I1204 11:43:34.961967 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:34Z","lastTransitionTime":"2025-12-04T11:43:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.064368 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.064432 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.064453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.064477 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.064496 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.167130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.167175 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.167186 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.167210 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.167222 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.269488 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.269542 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.269553 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.269572 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.269582 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.372644 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.372684 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.372693 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.372707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.372717 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.475855 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.475905 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.475919 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.475939 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.475952 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.578075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.578130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.578142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.578161 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.578174 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.681989 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.682102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.682122 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.682148 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.682164 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.785479 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.785580 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.785606 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.785640 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.785668 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.888786 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.888834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.888843 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.888859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.888868 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.991848 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.991889 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.991901 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.991919 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:35 crc kubenswrapper[4979]: I1204 11:43:35.991930 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:35Z","lastTransitionTime":"2025-12-04T11:43:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.094747 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.095219 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.095360 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.095448 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.095527 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198113 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198149 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198127 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:36 crc kubenswrapper[4979]: E1204 11:43:36.198246 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:36 crc kubenswrapper[4979]: E1204 11:43:36.198324 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:36 crc kubenswrapper[4979]: E1204 11:43:36.198516 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198894 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198924 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.198940 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.199225 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:36 crc kubenswrapper[4979]: E1204 11:43:36.199505 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.216076 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.230923 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.245998 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.261671 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.276580 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.293442 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.300770 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.300853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.300896 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.300917 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.300931 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.307925 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.321489 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.334980 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.347687 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.358377 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.371545 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.385688 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.400073 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:36 crc kubenswrapper[4979]: E1204 11:43:36.400260 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:36 crc kubenswrapper[4979]: E1204 11:43:36.400410 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:43:44.400385071 +0000 UTC m=+48.674680875 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.404579 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.404766 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.404797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.404807 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.404827 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.404838 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.424660 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.440621 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.453423 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.507665 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.507714 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.507724 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.507744 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.507756 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.611077 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.611144 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.611157 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.611176 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.611190 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.714472 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.714530 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.714543 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.714567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.714579 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.817652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.817755 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.817769 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.817811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.817826 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.920884 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.920936 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.920952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.920973 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:36 crc kubenswrapper[4979]: I1204 11:43:36.920988 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:36Z","lastTransitionTime":"2025-12-04T11:43:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.024030 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.024086 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.024097 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.024117 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.024131 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.127489 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.127552 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.127568 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.127593 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.127609 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.150995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.151093 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.151146 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.151177 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.151222 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: E1204 11:43:37.167769 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:37Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.172189 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.172233 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.172248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.172270 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.172286 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: E1204 11:43:37.188648 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:37Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.193485 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.193545 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.193564 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.193587 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.193603 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: E1204 11:43:37.209721 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:37Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.215509 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.215563 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.215573 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.215591 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.215606 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: E1204 11:43:37.232909 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:37Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.237148 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.237214 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.237232 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.237260 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.237275 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: E1204 11:43:37.251705 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:37Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:37 crc kubenswrapper[4979]: E1204 11:43:37.251865 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.253969 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.254008 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.254024 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.254043 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.254056 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.357076 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.357154 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.357178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.357268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.357294 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.461014 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.461069 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.461087 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.461108 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.461122 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.563767 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.563846 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.563856 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.563878 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.563889 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.667239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.667296 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.667334 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.667354 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.667364 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.770274 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.770537 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.770549 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.770572 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.770584 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.874100 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.874205 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.874222 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.874256 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.874270 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.977368 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.977433 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.977447 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.977466 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:37 crc kubenswrapper[4979]: I1204 11:43:37.977477 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:37Z","lastTransitionTime":"2025-12-04T11:43:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.080979 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.081042 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.081056 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.081077 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.081094 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.184227 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.184293 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.184437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.184464 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.184481 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.198708 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.198785 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.198764 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.198745 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:38 crc kubenswrapper[4979]: E1204 11:43:38.198912 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:38 crc kubenswrapper[4979]: E1204 11:43:38.199514 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:38 crc kubenswrapper[4979]: E1204 11:43:38.199567 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:38 crc kubenswrapper[4979]: E1204 11:43:38.199720 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.287097 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.287164 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.287184 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.287209 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.287222 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.390241 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.390322 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.390334 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.390352 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.390362 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.493526 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.493593 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.493606 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.493627 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.493666 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.596794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.596859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.596872 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.596893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.596907 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.699445 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.699521 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.699546 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.699571 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.699589 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.802192 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.802251 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.802262 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.802282 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.802312 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.905272 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.905343 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.905354 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.905371 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:38 crc kubenswrapper[4979]: I1204 11:43:38.905381 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:38Z","lastTransitionTime":"2025-12-04T11:43:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.008496 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.008571 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.008584 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.008611 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.008628 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.111787 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.111832 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.111843 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.111860 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.111871 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.214623 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.214668 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.214679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.214699 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.214715 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.317611 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.317677 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.317694 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.317721 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.317737 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.420567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.420627 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.420637 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.420656 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.420667 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.524534 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.525125 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.525163 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.525187 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.525199 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.628637 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.628713 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.628727 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.628753 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.628769 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.731366 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.731408 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.731424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.731443 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.731454 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.834361 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.834393 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.834402 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.834418 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.834428 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.937194 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.937259 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.937275 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.937321 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:39 crc kubenswrapper[4979]: I1204 11:43:39.937335 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:39Z","lastTransitionTime":"2025-12-04T11:43:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.040344 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.040396 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.040407 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.040432 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.040445 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.143596 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.143660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.143675 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.143698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.143714 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.198222 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.198383 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:40 crc kubenswrapper[4979]: E1204 11:43:40.198455 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.198268 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.198501 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:40 crc kubenswrapper[4979]: E1204 11:43:40.198657 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:40 crc kubenswrapper[4979]: E1204 11:43:40.198777 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:40 crc kubenswrapper[4979]: E1204 11:43:40.198919 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.247021 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.247094 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.247110 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.247128 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.247141 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.349543 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.349593 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.349607 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.349626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.349639 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.452733 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.452802 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.452814 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.452834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.452850 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.555457 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.555516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.555539 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.555560 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.555573 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.657723 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.657781 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.657794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.657811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.657823 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.760532 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.760605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.760615 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.760641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.760662 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.863282 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.863358 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.863370 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.863391 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.863404 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.966063 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.966128 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.966151 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.966181 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:40 crc kubenswrapper[4979]: I1204 11:43:40.966204 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:40Z","lastTransitionTime":"2025-12-04T11:43:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.069654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.069714 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.069730 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.069751 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.069766 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.172654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.172756 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.172771 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.172793 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.172809 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.276064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.276108 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.276118 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.276132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.276141 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.378706 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.378752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.378764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.378780 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.378821 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.481950 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.482017 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.482035 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.482058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.482075 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.585633 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.585676 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.585687 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.585704 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.585716 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.688900 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.688962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.688979 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.689013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.689029 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.791944 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.791985 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.791995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.792011 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.792021 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.895456 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.895521 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.895536 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.895554 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.895569 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.998791 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.998838 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.998850 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.998871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:41 crc kubenswrapper[4979]: I1204 11:43:41.998883 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:41Z","lastTransitionTime":"2025-12-04T11:43:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.102357 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.102399 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.102409 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.102424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.102434 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.199026 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.199114 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.199147 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.199213 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:42 crc kubenswrapper[4979]: E1204 11:43:42.199346 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:42 crc kubenswrapper[4979]: E1204 11:43:42.199460 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:42 crc kubenswrapper[4979]: E1204 11:43:42.199605 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:42 crc kubenswrapper[4979]: E1204 11:43:42.199827 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.204169 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.204203 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.204213 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.204229 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.204238 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.306995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.307041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.307052 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.307069 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.307079 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.410767 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.410842 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.410864 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.410889 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.410907 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.513974 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.514015 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.514026 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.514046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.514057 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.617714 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.617792 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.617805 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.617826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.617838 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.720573 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.720611 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.720620 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.720636 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.720649 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.823879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.823931 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.823949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.823974 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.823989 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.927168 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.927231 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.927249 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.927273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:42 crc kubenswrapper[4979]: I1204 11:43:42.927322 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:42Z","lastTransitionTime":"2025-12-04T11:43:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.029871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.029908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.029919 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.029936 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.029947 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.132475 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.132520 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.132530 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.132558 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.132568 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.235648 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.235699 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.235715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.235734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.235748 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.337841 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.337908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.337918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.337932 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.337943 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.440775 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.440834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.440847 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.440864 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.440877 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.543711 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.543764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.543779 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.543799 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.543812 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.646016 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.646084 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.646098 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.646122 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.646135 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.749437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.749495 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.749507 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.749717 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.749731 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.853976 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.854032 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.854051 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.854075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.854086 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.956980 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.957034 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.957048 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.957067 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:43 crc kubenswrapper[4979]: I1204 11:43:43.957079 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:43Z","lastTransitionTime":"2025-12-04T11:43:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.060380 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.060432 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.060447 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.060465 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.060479 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.163831 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.163958 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.163971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.163993 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.164072 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.198082 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.198196 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.198241 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:44 crc kubenswrapper[4979]: E1204 11:43:44.198442 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.198487 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:44 crc kubenswrapper[4979]: E1204 11:43:44.198573 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:44 crc kubenswrapper[4979]: E1204 11:43:44.198708 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:44 crc kubenswrapper[4979]: E1204 11:43:44.198846 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.266957 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.267007 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.267019 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.267035 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.267046 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.370674 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.370848 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.370875 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.370963 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.371051 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.474456 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.474520 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.474535 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.474557 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.474568 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.489208 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:44 crc kubenswrapper[4979]: E1204 11:43:44.489545 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:44 crc kubenswrapper[4979]: E1204 11:43:44.489634 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:44:00.489604239 +0000 UTC m=+64.763900043 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.535012 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.547375 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.553838 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.568082 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.578003 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.578138 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.578934 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.579002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.579112 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.584351 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.596974 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.609593 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.624127 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.636946 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.655680 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.671026 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.682491 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.682522 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.682531 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.682545 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.682556 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.683879 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.702703 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.717195 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.729863 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.743537 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.764749 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.786592 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.786646 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.786854 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.786876 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.786892 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.789008 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.803810 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:44Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.890705 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.890766 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.890785 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.890811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.890830 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.994756 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.994802 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.994814 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.994837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:44 crc kubenswrapper[4979]: I1204 11:43:44.994863 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:44Z","lastTransitionTime":"2025-12-04T11:43:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.097871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.097924 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.097941 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.097962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.097974 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.200446 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.200866 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.200962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.201072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.201164 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.305074 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.305123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.305135 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.305155 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.305169 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.407908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.407959 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.407969 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.407987 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.408000 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.510857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.511291 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.511315 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.511332 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.511346 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.614578 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.614651 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.614666 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.614683 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.614698 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.717530 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.717576 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.717591 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.717610 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.717625 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.820990 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.821040 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.821049 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.821065 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.821079 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.923867 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.923949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.924013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.924047 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:45 crc kubenswrapper[4979]: I1204 11:43:45.924075 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:45Z","lastTransitionTime":"2025-12-04T11:43:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.026961 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.027006 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.027015 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.027033 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.027045 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.107157 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.107373 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.107434 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.107459 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.107484 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:44:18.107441451 +0000 UTC m=+82.381737255 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.107533 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:44:18.107523203 +0000 UTC m=+82.381819007 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.107595 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.107648 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:44:18.107632556 +0000 UTC m=+82.381928360 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.130252 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.130313 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.130323 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.130342 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.130354 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.198729 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.198784 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.198927 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.198973 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.200549 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.200655 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.200736 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.200809 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.201083 4979 scope.go:117] "RemoveContainer" containerID="e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.208539 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.208610 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208748 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208784 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208800 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208860 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:44:18.208842758 +0000 UTC m=+82.483138562 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208854 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208913 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.208934 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:46 crc kubenswrapper[4979]: E1204 11:43:46.209037 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:44:18.209004582 +0000 UTC m=+82.483300556 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.216183 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.234017 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.234574 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.234627 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.234637 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.234657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.234670 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.250831 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.266153 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.282913 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.297485 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.310597 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.325086 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.336742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.336797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.336806 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.336821 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.336853 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.338607 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.349723 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.363488 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.378638 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.395263 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.407415 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.433429 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.440284 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.440324 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.440334 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.440351 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.440361 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.454853 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.472194 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.487999 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:46Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.542899 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.542950 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.542961 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.542980 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.542991 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.646095 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.646224 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.646252 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.646285 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.646428 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.748733 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.748781 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.748794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.748815 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.748828 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.851853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.851920 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.851932 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.851956 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.851968 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.954374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.954414 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.954427 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.954445 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:46 crc kubenswrapper[4979]: I1204 11:43:46.954457 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:46Z","lastTransitionTime":"2025-12-04T11:43:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.057052 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.057094 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.057107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.057162 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.057186 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.160004 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.160094 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.160123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.160158 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.160182 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.263723 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.263797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.263817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.263848 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.263866 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.330480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.330535 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.330552 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.330574 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.330589 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: E1204 11:43:47.350788 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.355551 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.355612 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.355628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.355654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.355670 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: E1204 11:43:47.370014 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.377074 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.377132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.377143 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.377166 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.377177 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: E1204 11:43:47.395065 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.402213 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.402259 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.402271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.402292 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.402322 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: E1204 11:43:47.416004 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.419762 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.419796 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.419806 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.419824 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.419836 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: E1204 11:43:47.433128 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: E1204 11:43:47.433438 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.435290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.435365 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.435379 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.435400 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.435415 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.538020 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.538079 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.538092 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.538129 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.538141 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.640957 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.641023 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.641048 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.641077 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.641100 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.666438 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/1.log" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.669514 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.670570 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.689464 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.704234 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.718019 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.729067 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.751395 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.751477 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.751556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.751586 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.751609 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.751693 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.765562 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.779713 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.792437 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.812542 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.835911 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.847065 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.854329 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.854391 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.854406 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.854725 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.854998 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.861352 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.877556 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.892953 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.907031 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.920861 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.932795 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.946366 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:47Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.957853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.958053 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.958078 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.958106 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:47 crc kubenswrapper[4979]: I1204 11:43:47.958116 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:47Z","lastTransitionTime":"2025-12-04T11:43:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.060802 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.060836 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.060844 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.060859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.060870 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.164798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.164837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.164849 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.164865 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.164878 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.199104 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:48 crc kubenswrapper[4979]: E1204 11:43:48.199290 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.199287 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:48 crc kubenswrapper[4979]: E1204 11:43:48.199697 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.199884 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:48 crc kubenswrapper[4979]: E1204 11:43:48.199964 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.200083 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:48 crc kubenswrapper[4979]: E1204 11:43:48.200266 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.267869 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.267922 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.267935 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.267955 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.267967 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.371290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.371369 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.371385 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.371405 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.371418 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.474570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.474623 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.474635 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.474651 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.474669 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.578358 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.578437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.578455 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.578479 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.578494 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.676219 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/2.log" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.677184 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/1.log" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680224 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680240 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8" exitCode=1 Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680265 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680279 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680319 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680334 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680275 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680462 4979 scope.go:117] "RemoveContainer" containerID="e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.680991 4979 scope.go:117] "RemoveContainer" containerID="644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8" Dec 04 11:43:48 crc kubenswrapper[4979]: E1204 11:43:48.681139 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.699140 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.715265 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.729506 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.743494 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.755186 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.769273 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.784842 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.785387 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.785405 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.785430 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.785451 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.788046 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.801111 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.814033 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.828135 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.844999 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.858831 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.872240 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.883739 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.888508 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.888570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.888590 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.888625 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.888642 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.907559 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.928451 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e95dfa5a68e5042dc372e8406eafe066c2ab36de279db762dfb4bc380350d177\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:26Z\\\",\\\"message\\\":\\\" Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.138\\\\\\\", Port:50051, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1204 11:43:25.381909 6413 services_controller.go:452] Built service openshift-marketplace/redhat-operators per-node LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381919 6413 services_controller.go:453] Built service openshift-marketplace/redhat-operators template LB for network=default: []services.LB{}\\\\nI1204 11:43:25.381923 6413 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:25.381930 6413 services_controller.go:454] Service openshift-marketplace/redhat-operators for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI1204 11:43:25.381776 6413 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1204 11:43:25.381989 6413 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.941511 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.954632 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:48Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.991172 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.991243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.991257 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.991279 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:48 crc kubenswrapper[4979]: I1204 11:43:48.991713 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:48Z","lastTransitionTime":"2025-12-04T11:43:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.094808 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.094852 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.094861 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.094877 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.094888 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.197691 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.197743 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.197755 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.197771 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.197781 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.300646 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.300913 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.300927 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.300948 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.300961 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.403857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.403915 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.403925 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.403944 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.403953 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.506879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.506954 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.506971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.506996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.507015 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.610420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.610479 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.610490 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.610509 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.610522 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.687893 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/2.log" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.698115 4979 scope.go:117] "RemoveContainer" containerID="644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8" Dec 04 11:43:49 crc kubenswrapper[4979]: E1204 11:43:49.698752 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.717776 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.717840 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.717864 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.717895 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.717918 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.719860 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.732882 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.755437 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.778819 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.793102 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.805228 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.820154 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.821127 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.821323 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.821465 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.821669 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.821867 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.832611 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.845046 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.855776 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.868808 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.882564 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.893862 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.904204 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.916756 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.929840 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.929897 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.929915 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.929934 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.929946 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:49Z","lastTransitionTime":"2025-12-04T11:43:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.934282 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.946424 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:49 crc kubenswrapper[4979]: I1204 11:43:49.965423 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:49Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.032900 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.033351 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.033472 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.033583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.033703 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.136740 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.136791 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.136802 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.136820 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.136834 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.199125 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:50 crc kubenswrapper[4979]: E1204 11:43:50.199400 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.199808 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.199828 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.199805 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:50 crc kubenswrapper[4979]: E1204 11:43:50.199920 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:50 crc kubenswrapper[4979]: E1204 11:43:50.200095 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:50 crc kubenswrapper[4979]: E1204 11:43:50.200222 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.239493 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.239556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.239580 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.239604 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.239621 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.342955 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.343018 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.343031 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.343053 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.343069 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.445940 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.445998 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.446011 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.446034 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.446046 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.548471 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.548902 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.549051 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.549147 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.549225 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.652637 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.652674 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.652683 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.652698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.652708 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.755994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.756045 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.756058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.756084 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.756098 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.859102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.859152 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.859165 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.859185 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.859197 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.962430 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.962526 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.962556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.962574 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:50 crc kubenswrapper[4979]: I1204 11:43:50.962587 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:50Z","lastTransitionTime":"2025-12-04T11:43:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.065605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.065654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.065667 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.065690 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.065701 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.168234 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.168290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.168330 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.168354 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.168369 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.271349 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.271398 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.271409 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.271431 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.271444 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.374725 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.374777 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.374790 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.374811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.374826 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.477879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.477934 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.477944 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.477965 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.477979 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.581446 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.581558 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.581573 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.581627 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.581647 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.684360 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.684413 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.684426 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.684444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.684457 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.787117 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.787179 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.787189 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.787208 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.787218 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.890162 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.890215 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.890225 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.890243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.890260 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.993191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.993263 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.993276 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.993327 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:51 crc kubenswrapper[4979]: I1204 11:43:51.993343 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:51Z","lastTransitionTime":"2025-12-04T11:43:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.096693 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.096742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.096752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.096771 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.096782 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.198039 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.198048 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:52 crc kubenswrapper[4979]: E1204 11:43:52.198245 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.198076 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.198048 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:52 crc kubenswrapper[4979]: E1204 11:43:52.198464 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:52 crc kubenswrapper[4979]: E1204 11:43:52.198374 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:52 crc kubenswrapper[4979]: E1204 11:43:52.198664 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.200416 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.200450 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.200459 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.200477 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.200488 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.303533 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.303575 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.303585 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.303603 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.303615 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.406708 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.407153 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.407332 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.407493 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.407639 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.511119 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.511178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.511192 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.511211 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.511223 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.613074 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.613118 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.613127 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.613146 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.613158 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.716225 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.716288 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.716369 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.716392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.716406 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.819786 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.819833 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.819842 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.819860 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.819871 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.921752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.921799 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.921809 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.921826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:52 crc kubenswrapper[4979]: I1204 11:43:52.921838 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:52Z","lastTransitionTime":"2025-12-04T11:43:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.025228 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.025352 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.025383 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.025410 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.025465 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.128426 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.128467 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.128480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.128497 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.128510 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.231350 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.231425 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.231497 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.231529 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.231556 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.334397 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.334429 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.334439 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.334454 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.334465 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.436809 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.436888 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.436907 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.436932 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.436950 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.539105 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.539144 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.539153 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.539168 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.539181 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.642473 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.642558 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.642583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.642612 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.642636 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.745476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.745516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.745524 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.745540 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.745551 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.849117 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.849173 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.849183 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.849199 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.849291 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.952914 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.952963 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.952974 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.952997 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:53 crc kubenswrapper[4979]: I1204 11:43:53.953012 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:53Z","lastTransitionTime":"2025-12-04T11:43:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.056595 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.056628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.056638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.056654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.056664 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.160129 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.160206 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.160224 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.160253 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.160272 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.198812 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.198969 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.199028 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.199061 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:54 crc kubenswrapper[4979]: E1204 11:43:54.199088 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:54 crc kubenswrapper[4979]: E1204 11:43:54.199135 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:54 crc kubenswrapper[4979]: E1204 11:43:54.199287 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:54 crc kubenswrapper[4979]: E1204 11:43:54.199403 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.262734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.262786 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.262795 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.262812 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.262823 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.366711 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.366789 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.366808 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.366839 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.366857 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.470170 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.470250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.470268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.470343 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.470367 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.573739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.573794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.573808 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.573832 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.573847 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.676817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.676870 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.676879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.676893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.676904 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.779484 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.779561 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.779577 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.779653 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.779667 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.882496 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.882550 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.882560 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.882579 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.882592 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.986916 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.986973 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.986988 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.987008 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:54 crc kubenswrapper[4979]: I1204 11:43:54.987022 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:54Z","lastTransitionTime":"2025-12-04T11:43:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.089942 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.090046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.090059 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.090107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.090120 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.193096 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.193147 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.193159 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.193181 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.193192 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.296598 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.296658 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.296670 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.296692 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.296702 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.399455 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.399522 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.399540 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.399566 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.399584 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.502343 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.502410 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.502423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.502449 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.502466 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.610693 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.610748 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.610762 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.610789 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.610807 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.714174 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.714275 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.714286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.714322 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.714335 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.817286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.817539 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.817558 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.817581 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.817601 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.920155 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.920243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.920261 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.920286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:55 crc kubenswrapper[4979]: I1204 11:43:55.920325 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:55Z","lastTransitionTime":"2025-12-04T11:43:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.022809 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.022861 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.022877 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.022898 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.022914 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.126352 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.126407 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.126422 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.126444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.126460 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.198412 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.198410 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.198553 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.198567 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:56 crc kubenswrapper[4979]: E1204 11:43:56.198674 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:56 crc kubenswrapper[4979]: E1204 11:43:56.198790 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:56 crc kubenswrapper[4979]: E1204 11:43:56.198933 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:56 crc kubenswrapper[4979]: E1204 11:43:56.198989 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.217399 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.228210 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.228246 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.228256 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.228273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.228287 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.233583 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.245972 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.258526 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.270402 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.286889 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.300422 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.314814 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.328484 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.330441 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.330490 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.330500 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.330513 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.330524 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.341132 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.352225 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.370465 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.386290 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.399490 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.423161 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.433341 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.433674 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.433745 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.433816 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.433891 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.443520 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.457380 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.470869 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:56Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.535944 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.535985 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.535997 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.536016 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.536028 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.638706 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.638742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.638751 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.638764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.638773 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.741631 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.741699 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.741718 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.741748 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.741767 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.844973 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.845453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.845607 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.845831 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.846012 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.949351 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.949400 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.949413 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.949434 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:56 crc kubenswrapper[4979]: I1204 11:43:56.949447 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:56Z","lastTransitionTime":"2025-12-04T11:43:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.052167 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.052221 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.052237 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.052261 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.052278 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.154857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.154892 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.154903 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.154916 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.154928 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.258267 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.258353 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.258364 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.258380 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.258391 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.361616 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.361668 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.361713 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.361795 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.361813 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.464663 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.464774 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.464798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.464827 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.464850 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.498276 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.498344 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.498357 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.498374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.498389 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: E1204 11:43:57.512821 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:57Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.517174 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.517239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.517251 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.517277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.517290 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: E1204 11:43:57.530935 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:57Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.535757 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.535838 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.535868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.535887 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.535898 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: E1204 11:43:57.548822 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:57Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.553041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.553098 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.553116 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.553145 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.553171 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: E1204 11:43:57.566415 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:57Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.571716 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.571777 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.571795 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.571825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.571845 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: E1204 11:43:57.585482 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:43:57Z is after 2025-08-24T17:21:41Z" Dec 04 11:43:57 crc kubenswrapper[4979]: E1204 11:43:57.585616 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.587647 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.587711 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.587733 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.587761 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.587783 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.690916 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.690953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.690963 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.690978 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.690989 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.795017 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.795064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.795073 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.795090 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.795103 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.897406 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.897462 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.897476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.897490 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:57 crc kubenswrapper[4979]: I1204 11:43:57.897502 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:57Z","lastTransitionTime":"2025-12-04T11:43:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.000994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.001068 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.001086 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.001112 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.001131 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.103880 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.103939 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.103952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.103972 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.103982 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.198818 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.198891 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.198945 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.198949 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:43:58 crc kubenswrapper[4979]: E1204 11:43:58.199000 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:43:58 crc kubenswrapper[4979]: E1204 11:43:58.199215 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:43:58 crc kubenswrapper[4979]: E1204 11:43:58.199335 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:43:58 crc kubenswrapper[4979]: E1204 11:43:58.199265 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.207552 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.207610 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.207623 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.207645 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.207658 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.310681 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.310731 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.310742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.310757 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.310768 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.413698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.413749 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.413764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.413787 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.413801 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.516504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.516546 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.516556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.516574 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.516586 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.619728 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.619780 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.619793 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.619812 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.619822 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.722513 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.722547 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.722556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.722573 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.722583 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.825783 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.825845 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.825857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.825877 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.825895 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.928629 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.928691 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.928707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.928729 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:58 crc kubenswrapper[4979]: I1204 11:43:58.928741 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:58Z","lastTransitionTime":"2025-12-04T11:43:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.031542 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.031608 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.031626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.031652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.031668 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.134830 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.134892 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.134903 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.134918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.134928 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.237428 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.237494 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.237502 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.237520 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.237530 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.340875 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.340937 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.340951 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.340975 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.340994 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.443680 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.443778 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.443802 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.443827 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.443848 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.547076 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.547119 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.547128 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.547143 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.547153 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.650570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.650601 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.650716 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.650735 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.650746 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.752906 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.752952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.752962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.752979 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.752990 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.856120 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.856183 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.856198 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.856222 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.856238 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.958641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.958679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.958687 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.958703 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:43:59 crc kubenswrapper[4979]: I1204 11:43:59.958714 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:43:59Z","lastTransitionTime":"2025-12-04T11:43:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.062002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.062063 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.062080 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.062100 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.062111 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.165066 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.165111 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.165124 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.165142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.165152 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.199061 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.199173 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.199073 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:00 crc kubenswrapper[4979]: E1204 11:44:00.199222 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.199090 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:00 crc kubenswrapper[4979]: E1204 11:44:00.199364 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:00 crc kubenswrapper[4979]: E1204 11:44:00.199528 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:00 crc kubenswrapper[4979]: E1204 11:44:00.199633 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.267969 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.268018 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.268031 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.268054 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.268073 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.371004 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.371088 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.371099 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.371132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.371143 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.473457 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.473499 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.473510 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.473529 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.473543 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.572642 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:00 crc kubenswrapper[4979]: E1204 11:44:00.572819 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:44:00 crc kubenswrapper[4979]: E1204 11:44:00.572947 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:44:32.572921265 +0000 UTC m=+96.847217069 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.576323 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.576355 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.576369 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.576386 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.576397 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.679382 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.679452 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.679473 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.679495 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.679518 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.782171 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.782218 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.782231 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.782250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.782264 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.885899 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.885972 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.885983 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.886005 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.886019 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.989709 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.989752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.989764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.989781 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:00 crc kubenswrapper[4979]: I1204 11:44:00.989792 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:00Z","lastTransitionTime":"2025-12-04T11:44:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.092385 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.092449 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.092460 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.092490 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.092504 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.199528 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.199607 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.199628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.199652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.199672 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.302947 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.303026 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.303041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.303072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.303097 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.406282 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.406410 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.406422 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.406439 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.406451 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.509493 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.509563 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.509581 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.509605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.509623 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.612612 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.612658 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.612671 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.612691 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.612707 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.715452 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.715500 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.715514 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.715534 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.715549 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.740862 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/0.log" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.740919 4979 generic.go:334] "Generic (PLEG): container finished" podID="12339390-77cf-4906-80c3-ddb729e42132" containerID="dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6" exitCode=1 Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.740960 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerDied","Data":"dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.741438 4979 scope.go:117] "RemoveContainer" containerID="dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.755811 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.773369 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.787552 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.802342 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.819208 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.819278 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.819288 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.819321 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.819335 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.820275 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.834769 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.847989 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.861429 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.872716 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.885787 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.896486 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.907349 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.918524 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.922024 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.922078 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.922090 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.922107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.922119 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:01Z","lastTransitionTime":"2025-12-04T11:44:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.929115 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.942394 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.954087 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.972990 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:01 crc kubenswrapper[4979]: I1204 11:44:01.991530 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:01Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.024740 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.024776 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.024784 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.024803 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.024822 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.128702 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.128745 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.128758 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.128776 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.128789 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.198785 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.198849 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.198857 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:02 crc kubenswrapper[4979]: E1204 11:44:02.198942 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.199016 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:02 crc kubenswrapper[4979]: E1204 11:44:02.199099 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:02 crc kubenswrapper[4979]: E1204 11:44:02.199205 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:02 crc kubenswrapper[4979]: E1204 11:44:02.199363 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.231630 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.231674 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.231687 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.231704 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.231716 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.335196 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.335246 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.335259 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.335283 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.335316 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.438721 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.438770 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.438781 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.438797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.438810 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.541877 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.541930 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.541946 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.541968 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.541983 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.644228 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.644273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.644285 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.644325 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.644337 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746058 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/0.log" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746146 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerStarted","Data":"40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746160 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746202 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746216 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.746226 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.761590 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.777970 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.792919 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.804783 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.816871 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.832090 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.846340 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.849385 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.849429 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.849452 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.849481 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.849506 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.860399 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.874587 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.895420 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.915102 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.925465 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.939781 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.950852 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.952096 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.952140 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.952153 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.952172 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.952186 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:02Z","lastTransitionTime":"2025-12-04T11:44:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.966955 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.980848 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:02 crc kubenswrapper[4979]: I1204 11:44:02.994963 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:02Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.009427 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:03Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.054268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.054338 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.054354 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.054377 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.054392 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.157223 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.157317 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.157331 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.157347 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.157357 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.260344 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.260393 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.260405 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.260423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.260437 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.363384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.363444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.363459 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.363481 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.363498 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.466645 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.466685 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.466698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.466715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.466728 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.569278 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.569346 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.569356 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.569374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.569385 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.672393 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.672445 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.672454 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.672474 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.672486 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.774871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.774924 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.774938 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.774956 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.774971 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.877423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.877471 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.877483 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.877500 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.877511 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.980454 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.980509 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.980525 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.980544 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:03 crc kubenswrapper[4979]: I1204 11:44:03.980557 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:03Z","lastTransitionTime":"2025-12-04T11:44:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.082989 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.083060 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.083072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.083090 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.083103 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.186392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.186460 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.186476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.186501 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.186521 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.198998 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:04 crc kubenswrapper[4979]: E1204 11:44:04.199214 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.199256 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.199279 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.199388 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:04 crc kubenswrapper[4979]: E1204 11:44:04.199935 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.200145 4979 scope.go:117] "RemoveContainer" containerID="644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8" Dec 04 11:44:04 crc kubenswrapper[4979]: E1204 11:44:04.200180 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:04 crc kubenswrapper[4979]: E1204 11:44:04.200332 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:04 crc kubenswrapper[4979]: E1204 11:44:04.200355 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.289829 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.289893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.289902 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.289921 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.289932 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.393293 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.393376 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.393388 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.393409 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.393423 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.496193 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.496239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.496248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.496265 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.496278 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.598658 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.598707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.598720 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.598739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.598755 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.702198 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.702258 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.702273 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.702294 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.702323 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.805199 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.805240 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.805248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.805262 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.805273 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.908191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.908250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.908284 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.908333 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:04 crc kubenswrapper[4979]: I1204 11:44:04.908346 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:04Z","lastTransitionTime":"2025-12-04T11:44:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.011128 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.011197 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.011215 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.011246 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.011264 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.114132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.114171 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.114181 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.114196 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.114208 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.217698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.217762 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.217775 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.217795 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.217810 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.320406 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.320447 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.320455 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.320471 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.320481 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.422985 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.423037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.423048 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.423066 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.423077 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.526124 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.526197 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.526214 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.526241 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.526261 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.629912 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.629971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.629983 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.630002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.630013 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.732178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.732239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.732248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.732263 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.732273 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.834209 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.834250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.834258 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.834276 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.834287 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.937453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.937496 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.937507 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.937523 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:05 crc kubenswrapper[4979]: I1204 11:44:05.937532 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:05Z","lastTransitionTime":"2025-12-04T11:44:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.039953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.040002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.040012 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.040032 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.040044 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.142098 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.142147 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.142158 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.142174 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.142183 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.198948 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.198999 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:06 crc kubenswrapper[4979]: E1204 11:44:06.199111 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.199133 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.198948 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:06 crc kubenswrapper[4979]: E1204 11:44:06.199276 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:06 crc kubenswrapper[4979]: E1204 11:44:06.199393 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:06 crc kubenswrapper[4979]: E1204 11:44:06.199502 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.212878 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.226496 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.241753 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.245225 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.245267 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.245277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.245293 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.245321 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.256230 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.272145 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.288960 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.302495 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.314469 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.329032 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.339140 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.348842 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.348897 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.348910 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.348931 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.348945 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.352309 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.365387 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.376053 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.386231 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.397155 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.416359 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.436611 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.446932 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:06Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.451871 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.451896 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.451904 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.451918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.451927 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.554150 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.554203 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.554215 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.554234 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.554247 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.657150 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.657195 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.657208 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.657225 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.657238 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.759134 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.759202 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.759219 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.759243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.759264 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.861723 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.861773 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.861787 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.861807 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.861818 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.965059 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.965125 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.965170 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.965195 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:06 crc kubenswrapper[4979]: I1204 11:44:06.965223 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:06Z","lastTransitionTime":"2025-12-04T11:44:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.068027 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.068072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.068083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.068102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.068114 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.170568 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.170640 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.170666 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.170706 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.170724 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.273225 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.273292 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.273341 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.273367 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.273388 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.376847 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.376907 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.376925 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.376953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.376972 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.479734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.479813 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.479838 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.479870 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.479892 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.582703 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.582761 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.582774 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.582799 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.582813 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.685742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.685823 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.685841 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.685865 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.685884 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.734815 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.735144 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.735254 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.735373 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.735473 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: E1204 11:44:07.756529 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:07Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.761817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.761880 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.761898 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.761917 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.761930 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: E1204 11:44:07.783794 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:07Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.789606 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.789796 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.789882 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.790008 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.790138 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: E1204 11:44:07.808459 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:07Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.813382 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.813437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.813450 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.813468 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.813480 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: E1204 11:44:07.828634 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:07Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.833137 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.833196 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.833209 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.833228 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.833240 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: E1204 11:44:07.846336 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:07Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:07 crc kubenswrapper[4979]: E1204 11:44:07.846595 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.848599 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.848632 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.848645 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.848664 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.848675 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.951353 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.951397 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.951408 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.951426 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:07 crc kubenswrapper[4979]: I1204 11:44:07.951437 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:07Z","lastTransitionTime":"2025-12-04T11:44:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.054411 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.054490 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.054501 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.054524 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.054536 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.157913 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.157972 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.157992 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.158025 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.158053 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.198797 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:08 crc kubenswrapper[4979]: E1204 11:44:08.198948 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.198994 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.199133 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:08 crc kubenswrapper[4979]: E1204 11:44:08.199461 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.199718 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:08 crc kubenswrapper[4979]: E1204 11:44:08.199830 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:08 crc kubenswrapper[4979]: E1204 11:44:08.200014 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.260423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.260475 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.260494 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.260517 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.260535 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.362817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.362870 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.362883 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.362902 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.362912 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.465219 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.465258 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.465272 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.465402 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.465440 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.568594 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.568659 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.568679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.568705 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.568725 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.671655 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.671701 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.671712 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.671730 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.671742 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.774167 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.774247 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.774275 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.774332 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.774354 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.876443 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.876500 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.876556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.876579 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.876595 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.979953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.980010 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.980023 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.980047 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:08 crc kubenswrapper[4979]: I1204 11:44:08.980061 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:08Z","lastTransitionTime":"2025-12-04T11:44:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.082862 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.082930 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.082949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.082977 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.082998 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.186647 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.186717 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.186736 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.186760 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.186779 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.290145 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.290204 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.290214 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.290235 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.290246 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.392987 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.393031 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.393041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.393058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.393071 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.496124 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.496165 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.496176 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.496194 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.496204 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.600019 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.600066 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.600077 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.600096 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.600109 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.703080 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.703134 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.703158 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.703172 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.703182 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.806456 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.806574 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.806592 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.806614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.806631 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.909614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.909679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.909696 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.909726 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:09 crc kubenswrapper[4979]: I1204 11:44:09.909750 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:09Z","lastTransitionTime":"2025-12-04T11:44:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.012778 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.012856 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.012878 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.012911 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.012933 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.115543 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.115593 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.115608 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.115627 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.115642 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.198476 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:10 crc kubenswrapper[4979]: E1204 11:44:10.198751 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.198822 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.198822 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.199001 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:10 crc kubenswrapper[4979]: E1204 11:44:10.199002 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:10 crc kubenswrapper[4979]: E1204 11:44:10.199115 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:10 crc kubenswrapper[4979]: E1204 11:44:10.199261 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.217799 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.217837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.217853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.217868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.217878 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.320051 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.320098 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.320107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.320135 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.320145 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.423286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.423351 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.423360 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.423375 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.423384 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.526892 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.526975 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.526999 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.527028 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.527049 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.630969 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.631041 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.631059 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.631086 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.631107 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.735195 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.735248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.735262 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.735280 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.735291 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.839754 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.839844 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.839864 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.839890 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.839912 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.944417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.944465 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.944480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.944502 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:10 crc kubenswrapper[4979]: I1204 11:44:10.944516 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:10Z","lastTransitionTime":"2025-12-04T11:44:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.046843 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.046890 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.046902 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.046921 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.046936 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.150150 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.150191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.150204 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.150223 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.150236 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.252791 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.252844 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.252856 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.252875 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.252889 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.356816 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.356876 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.356893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.356918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.356936 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.460288 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.460377 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.460401 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.460425 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.460441 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.562860 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.562939 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.562952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.562977 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.562990 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.665769 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.665836 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.665849 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.665927 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.665953 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.770960 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.771037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.771056 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.771087 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.771105 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.874198 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.874290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.874383 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.874423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.874503 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.977911 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.978074 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.978096 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.978122 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:11 crc kubenswrapper[4979]: I1204 11:44:11.978139 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:11Z","lastTransitionTime":"2025-12-04T11:44:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.081848 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.081925 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.081943 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.081964 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.081977 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.185288 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.185366 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.185381 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.185401 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.185415 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.199057 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.199108 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.199090 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.199055 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:12 crc kubenswrapper[4979]: E1204 11:44:12.199235 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:12 crc kubenswrapper[4979]: E1204 11:44:12.199384 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:12 crc kubenswrapper[4979]: E1204 11:44:12.199514 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:12 crc kubenswrapper[4979]: E1204 11:44:12.199562 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.288551 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.288603 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.288614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.288677 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.288693 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.391788 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.391860 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.391878 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.391904 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.391922 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.494660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.494958 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.495043 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.495123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.495229 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.598092 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.598690 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.598944 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.599142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.599278 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.702183 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.702221 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.702233 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.702249 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.702261 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.805889 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.806319 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.806417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.806510 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.806585 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.909718 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.909773 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.909784 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.909799 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:12 crc kubenswrapper[4979]: I1204 11:44:12.909819 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:12Z","lastTransitionTime":"2025-12-04T11:44:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.012024 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.012099 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.012124 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.012154 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.012176 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.114907 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.114978 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.114996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.115023 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.115042 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.218628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.218732 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.218752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.218811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.218830 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.321814 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.321906 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.321963 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.321989 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.322007 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.425085 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.425137 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.425157 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.425177 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.425195 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.528632 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.529196 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.529348 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.529449 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.529543 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.632665 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.632736 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.632749 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.632794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.632809 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.735940 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.735986 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.735997 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.736014 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.736024 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.839686 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.839772 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.839794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.839821 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.839841 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.943384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.943434 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.943443 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.943458 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:13 crc kubenswrapper[4979]: I1204 11:44:13.943470 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:13Z","lastTransitionTime":"2025-12-04T11:44:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.047071 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.047153 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.047179 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.047211 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.047235 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.149690 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.149740 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.149752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.149777 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.149795 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.198484 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.198534 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.198547 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:14 crc kubenswrapper[4979]: E1204 11:44:14.198670 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.198711 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:14 crc kubenswrapper[4979]: E1204 11:44:14.198868 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:14 crc kubenswrapper[4979]: E1204 11:44:14.199000 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:14 crc kubenswrapper[4979]: E1204 11:44:14.199131 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.252504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.252559 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.252583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.252606 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.252623 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.355959 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.356004 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.356016 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.356035 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.356045 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.458440 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.458497 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.458514 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.458537 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.458552 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.561104 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.561556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.561692 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.561817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.561994 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.665177 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.665253 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.665277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.665349 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.665376 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.769210 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.769668 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.769762 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.769832 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.769906 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.872894 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.872952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.872970 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.872996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.873015 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.976994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.977044 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.977057 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.977075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:14 crc kubenswrapper[4979]: I1204 11:44:14.977086 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:14Z","lastTransitionTime":"2025-12-04T11:44:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.079652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.079707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.079718 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.079737 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.079748 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.182338 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.182423 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.182444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.182473 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.182493 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.214546 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.286437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.286491 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.286504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.286527 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.286540 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.390375 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.391076 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.391277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.391472 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.391599 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.494468 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.494819 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.494910 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.495121 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.495363 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.597879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.597951 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.597967 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.597984 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.597997 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.702214 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.702286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.702342 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.702373 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.702391 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.804798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.804865 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.804881 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.804904 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.804918 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.907763 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.908185 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.908381 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.908540 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:15 crc kubenswrapper[4979]: I1204 11:44:15.908641 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:15Z","lastTransitionTime":"2025-12-04T11:44:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.012289 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.012374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.012391 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.012419 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.012437 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.115894 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.115959 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.116137 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.116192 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.116222 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.198709 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.198726 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.198788 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.198882 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:16 crc kubenswrapper[4979]: E1204 11:44:16.199268 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:16 crc kubenswrapper[4979]: E1204 11:44:16.199377 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:16 crc kubenswrapper[4979]: E1204 11:44:16.199177 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.199467 4979 scope.go:117] "RemoveContainer" containerID="644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8" Dec 04 11:44:16 crc kubenswrapper[4979]: E1204 11:44:16.199514 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.218225 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.219831 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.219899 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.219913 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.219934 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.219949 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.234164 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.263735 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.301764 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.319519 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.322020 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.322073 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.322083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.322118 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.322131 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.332327 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.345411 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.361391 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.376139 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.391860 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.414686 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.425083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.425117 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.425126 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.425139 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.425149 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.429216 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.439853 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.451211 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.463526 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.473576 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b87b21ad-7aaa-443d-b9e5-984c87d2d20f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.482926 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.494546 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.508084 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.528221 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.528267 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.528275 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.528293 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.528325 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.631424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.631494 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.631510 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.631557 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.631571 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.734598 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.734657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.734673 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.734695 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.734711 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.803155 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/2.log" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.807947 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.808374 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.827123 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.838737 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.838795 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.838810 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.838831 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.838852 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.844029 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.860343 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.887277 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.914689 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.941048 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.942068 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.942108 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.942123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.942142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.942153 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:16Z","lastTransitionTime":"2025-12-04T11:44:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.959244 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:16 crc kubenswrapper[4979]: I1204 11:44:16.977365 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.000843 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:16Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.016999 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.029658 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.044417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.044454 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.044462 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.044476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.044487 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.045003 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.056013 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.065927 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.076820 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.087976 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.096974 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b87b21ad-7aaa-443d-b9e5-984c87d2d20f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.109350 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.121817 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.146860 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.146910 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.146921 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.146943 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.146981 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.250211 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.250267 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.250278 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.250311 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.250325 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.356563 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.356615 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.356638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.356664 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.356678 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.459861 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.459902 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.459911 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.459925 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.459935 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.562654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.562700 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.562712 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.562731 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.562744 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.665993 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.666067 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.666092 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.666121 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.666141 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.769533 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.769608 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.769632 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.769668 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.769695 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.815137 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/3.log" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.816279 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/2.log" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.820936 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" exitCode=1 Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.820984 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.821070 4979 scope.go:117] "RemoveContainer" containerID="644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.822176 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:44:17 crc kubenswrapper[4979]: E1204 11:44:17.822825 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.838025 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b87b21ad-7aaa-443d-b9e5-984c87d2d20f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.853060 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.868418 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.873253 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.873373 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.873397 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.873425 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.873443 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.881419 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.890863 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.902512 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.912737 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.926108 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.937617 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.960069 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.976551 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.977118 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.977132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.977151 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.977167 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:17Z","lastTransitionTime":"2025-12-04T11:44:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.982043 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://644f96b8836828e8f7834c83a4361c292fe95bb0c240ff3a01ae6bd0d021f1a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:43:48Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1204 11:43:47.663393 6668 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1204 11:43:47.663178 6668 default_network_controller.go:776] Recording success event on pod openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c\\\\nI1204 11:43:47.663046 6668 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1204 11:43:47.663200 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663447 6668 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-x7qhm\\\\nI1204 11:43:47.663472 6668 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-x7qhm in node crc\\\\nI1204 11:43:47.663221 6668 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-target-xd92c\\\\nF1204 11:43:47.663507 6668 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:17Z\\\",\\\"message\\\":\\\"ba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z]\\\\nI1204 11:44:17.379359 7043 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:44:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:17 crc kubenswrapper[4979]: I1204 11:44:17.995921 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.006637 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.012004 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.012052 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.012062 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.012079 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.012091 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.024929 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.026088 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.029853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.029927 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.029946 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.029975 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.030004 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.043743 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.044164 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.047828 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.047864 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.047878 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.047897 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.047908 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.059938 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.061603 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.070075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.070132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.070149 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.070171 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.070184 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.082145 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.082329 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.087569 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.087652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.087669 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.087689 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.087708 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.103938 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.111936 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.112072 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.114400 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.114532 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.114633 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.114733 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.114817 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.120512 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.168334 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.168487 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.168538 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.168501884 +0000 UTC m=+146.442797688 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.168565 4979 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.168633 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.168614687 +0000 UTC m=+146.442910491 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.168688 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.168817 4979 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.168869 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.168862254 +0000 UTC m=+146.443158058 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.198249 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.198389 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.198382 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.198382 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.198511 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.198707 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.198939 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.199115 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.217908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.217947 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.217961 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.217980 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.217996 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.270157 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.270227 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270437 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270475 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270482 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270496 4979 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270506 4979 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270525 4979 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270573 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.270546478 +0000 UTC m=+146.544842292 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.270596 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.270587589 +0000 UTC m=+146.544883403 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.320403 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.320826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.320917 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.321018 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.321113 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.423788 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.423834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.423843 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.423857 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.423871 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.527759 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.527870 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.527883 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.527901 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.527915 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.631365 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.631430 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.631454 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.631486 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.631512 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.735144 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.735246 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.735274 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.735353 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.735387 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.827580 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/3.log" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.831882 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:44:18 crc kubenswrapper[4979]: E1204 11:44:18.833360 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.838338 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.838396 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.838407 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.838426 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.838441 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.850394 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b87b21ad-7aaa-443d-b9e5-984c87d2d20f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.864435 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.880111 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.898457 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.914872 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.937394 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.942804 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.942844 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.942859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.942879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.942896 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:18Z","lastTransitionTime":"2025-12-04T11:44:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.949570 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.962049 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.974769 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:18 crc kubenswrapper[4979]: I1204 11:44:18.994643 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:18Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.019176 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:17Z\\\",\\\"message\\\":\\\"ba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z]\\\\nI1204 11:44:17.379359 7043 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:44:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.034973 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.046002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.046059 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.046071 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.046087 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.046100 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.050369 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.064131 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.080809 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.094096 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.107417 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.124976 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.143692 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:19Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.148463 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.148585 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.148658 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.148734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.148807 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.251581 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.251630 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.251641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.251659 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.251673 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.355060 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.355546 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.355656 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.355763 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.355838 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.459027 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.459084 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.459105 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.459134 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.459153 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.561649 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.561700 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.561714 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.561731 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.561741 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.665173 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.665224 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.665237 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.665255 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.665269 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.767924 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.767988 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.768003 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.768030 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.768049 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.871163 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.871230 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.871251 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.871279 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.871329 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.974518 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.974996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.975087 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.975169 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:19 crc kubenswrapper[4979]: I1204 11:44:19.975243 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:19Z","lastTransitionTime":"2025-12-04T11:44:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.077702 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.077758 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.077768 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.077786 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.077796 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.180726 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.181115 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.181241 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.181386 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.181543 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.198744 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.198797 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.198838 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.198814 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:20 crc kubenswrapper[4979]: E1204 11:44:20.198934 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:20 crc kubenswrapper[4979]: E1204 11:44:20.199015 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:20 crc kubenswrapper[4979]: E1204 11:44:20.199092 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:20 crc kubenswrapper[4979]: E1204 11:44:20.199135 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.284603 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.284653 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.284664 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.284683 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.284694 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.388782 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.388868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.388892 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.388928 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.388956 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.492281 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.492748 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.492835 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.492933 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.493046 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.595779 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.595859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.595878 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.595904 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.595923 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.699778 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.699816 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.699828 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.699847 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.699864 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.802745 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.803544 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.803739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.803888 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.804010 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.907587 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.908044 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.908271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.908416 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:20 crc kubenswrapper[4979]: I1204 11:44:20.908698 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:20Z","lastTransitionTime":"2025-12-04T11:44:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.012255 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.012797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.013056 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.013218 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.013387 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.116187 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.116240 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.116257 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.116282 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.116328 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.219804 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.220257 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.220378 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.220461 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.220521 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.323415 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.323476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.323491 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.323512 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.323529 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.426749 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.426794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.426806 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.426824 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.426839 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.530028 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.530617 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.530773 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.530908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.531024 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.635089 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.635158 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.635181 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.635210 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.635232 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.737990 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.738477 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.738589 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.738691 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.738795 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.841599 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.841641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.841652 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.841671 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.841686 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.945130 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.945621 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.945865 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.946126 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:21 crc kubenswrapper[4979]: I1204 11:44:21.946294 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:21Z","lastTransitionTime":"2025-12-04T11:44:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.050748 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.050797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.050813 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.050836 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.050853 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.154188 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.154261 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.154275 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.154325 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.154341 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.199094 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.199327 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.199291 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.199371 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:22 crc kubenswrapper[4979]: E1204 11:44:22.199434 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:22 crc kubenswrapper[4979]: E1204 11:44:22.199470 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:22 crc kubenswrapper[4979]: E1204 11:44:22.202444 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:22 crc kubenswrapper[4979]: E1204 11:44:22.202857 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.257452 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.257520 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.257538 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.257557 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.257570 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.360540 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.360605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.360623 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.360657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.360681 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.463713 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.463785 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.463822 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.463853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.463873 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.566883 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.567373 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.567641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.567847 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.567994 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.671984 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.672018 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.672029 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.672046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.672057 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.775254 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.775320 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.775337 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.775353 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.775366 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.878669 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.878716 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.878733 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.878750 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.878763 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.982257 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.982369 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.982386 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.982413 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:22 crc kubenswrapper[4979]: I1204 11:44:22.982430 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:22Z","lastTransitionTime":"2025-12-04T11:44:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.085595 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.085649 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.085661 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.085682 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.085699 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.188796 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.188868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.188884 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.188905 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.188918 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.292533 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.292585 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.292597 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.292619 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.292630 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.395428 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.395503 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.395527 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.395555 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.395569 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.498738 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.498791 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.498805 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.498826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.498845 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.602516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.602587 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.602605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.602633 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.602654 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.709363 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.709419 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.709434 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.709460 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.709482 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.812440 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.812516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.812535 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.812564 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.812583 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.915707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.915801 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.915824 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.915859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:23 crc kubenswrapper[4979]: I1204 11:44:23.915885 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:23Z","lastTransitionTime":"2025-12-04T11:44:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.019598 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.019670 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.019693 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.019722 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.019741 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.122786 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.123007 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.123057 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.123101 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.123126 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.198779 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.198842 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.198780 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:24 crc kubenswrapper[4979]: E1204 11:44:24.199003 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:24 crc kubenswrapper[4979]: E1204 11:44:24.199132 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.199194 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:24 crc kubenswrapper[4979]: E1204 11:44:24.199279 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:24 crc kubenswrapper[4979]: E1204 11:44:24.199467 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.226038 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.226114 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.226132 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.226164 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.226183 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.329675 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.329758 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.329788 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.329824 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.329845 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.434362 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.434426 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.434441 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.434464 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.434482 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.536801 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.536868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.536888 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.536915 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.536933 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.640348 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.640394 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.640404 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.640424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.640435 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.742933 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.742993 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.743009 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.743030 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.743046 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.854971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.855057 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.855073 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.855093 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.855108 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.958434 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.958519 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.958534 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.958556 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:24 crc kubenswrapper[4979]: I1204 11:44:24.958570 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:24Z","lastTransitionTime":"2025-12-04T11:44:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.061118 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.061167 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.061181 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.061200 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.061212 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.163567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.163631 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.163647 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.163667 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.163681 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.266633 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.266678 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.266686 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.266700 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.266709 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.369628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.369704 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.369717 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.369743 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.369755 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.472794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.472862 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.472882 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.472918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.472955 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.575200 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.575235 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.575246 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.575262 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.575272 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.678775 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.678859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.678882 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.678914 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.678935 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.782133 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.782173 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.782191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.782215 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.782234 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.885689 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.886178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.886206 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.886265 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.886283 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.989585 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.989638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.989651 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.989669 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:25 crc kubenswrapper[4979]: I1204 11:44:25.989680 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:25Z","lastTransitionTime":"2025-12-04T11:44:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.092996 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.093064 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.093076 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.093098 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.093111 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.195577 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.195628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.195640 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.195660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.195673 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.198202 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.198254 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.198227 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:26 crc kubenswrapper[4979]: E1204 11:44:26.198468 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:26 crc kubenswrapper[4979]: E1204 11:44:26.198567 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.198622 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:26 crc kubenswrapper[4979]: E1204 11:44:26.198753 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:26 crc kubenswrapper[4979]: E1204 11:44:26.198888 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.216666 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.231997 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.243945 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b87b21ad-7aaa-443d-b9e5-984c87d2d20f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.258180 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.277291 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.291555 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.297493 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.297544 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.297559 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.297581 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.297596 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.305054 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.322697 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.336812 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.360404 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.380851 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:17Z\\\",\\\"message\\\":\\\"ba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z]\\\\nI1204 11:44:17.379359 7043 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:44:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.395200 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.401127 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.401167 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.401177 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.401196 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.401209 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.409228 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.427765 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.441932 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.460592 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.477826 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.494048 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.504037 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.504092 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.504102 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.504120 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.504130 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.510045 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:26Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.607570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.607614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.607626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.607643 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.607653 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.710838 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.710895 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.710912 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.710941 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.710958 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.813621 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.813679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.813694 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.813826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.813847 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.918906 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.918947 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.918959 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.918978 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:26 crc kubenswrapper[4979]: I1204 11:44:26.918990 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:26Z","lastTransitionTime":"2025-12-04T11:44:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.021378 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.021920 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.022083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.022232 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.022385 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.126940 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.126990 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.127000 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.127023 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.127043 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.230075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.230129 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.230142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.230162 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.230176 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.333164 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.333233 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.333246 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.333268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.333281 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.437015 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.437075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.437088 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.437111 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.437126 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.540584 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.540663 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.540692 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.540765 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.540793 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.643939 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.644024 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.644050 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.644083 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.644105 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.746890 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.746947 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.746957 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.746973 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.747005 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.850458 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.850504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.850514 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.850530 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.850542 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.953471 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.953537 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.953547 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.953569 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:27 crc kubenswrapper[4979]: I1204 11:44:27.953581 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:27Z","lastTransitionTime":"2025-12-04T11:44:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.057117 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.057162 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.057176 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.057196 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.057209 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.159950 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.160013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.160027 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.160046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.160059 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.198824 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.198924 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.198848 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.198837 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.199022 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.199154 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.199271 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.199419 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.262646 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.262693 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.262704 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.262720 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.262730 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.365427 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.365489 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.365499 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.365514 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.365524 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.442142 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.442191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.442207 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.442231 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.442248 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.464860 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.470040 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.470113 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.470135 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.470161 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.470179 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.490768 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.497287 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.497347 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.497361 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.497382 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.497397 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.518343 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.525138 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.525286 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.525323 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.525350 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.525362 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.540970 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.546231 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.546326 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.546345 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.546371 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.546393 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.562536 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:28Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:28 crc kubenswrapper[4979]: E1204 11:44:28.562702 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.565070 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.565160 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.565218 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.565243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.565287 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.668158 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.668233 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.668249 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.668271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.668287 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.771810 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.771879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.771893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.771913 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.771928 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.874108 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.874169 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.874179 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.874193 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.874204 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.977679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.977730 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.977752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.977782 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:28 crc kubenswrapper[4979]: I1204 11:44:28.977803 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:28Z","lastTransitionTime":"2025-12-04T11:44:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.081178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.081243 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.081252 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.081270 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.081282 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.184419 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.184501 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.184521 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.184546 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.184563 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.288156 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.288206 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.288218 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.288238 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.288252 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.391901 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.391961 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.391973 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.391995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.392010 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.494834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.494900 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.494911 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.494931 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.494945 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.597539 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.597604 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.597617 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.597639 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.597653 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.700839 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.700893 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.700906 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.700956 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.700970 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.803849 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.803923 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.803940 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.803967 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.803986 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.906698 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.906739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.906748 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.906764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:29 crc kubenswrapper[4979]: I1204 11:44:29.906775 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:29Z","lastTransitionTime":"2025-12-04T11:44:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.009645 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.009707 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.009729 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.009752 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.009774 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.113076 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.113134 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.113152 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.113174 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.113186 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.198102 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.198135 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.198242 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:30 crc kubenswrapper[4979]: E1204 11:44:30.198543 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.198611 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:30 crc kubenswrapper[4979]: E1204 11:44:30.198786 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:30 crc kubenswrapper[4979]: E1204 11:44:30.198995 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:30 crc kubenswrapper[4979]: E1204 11:44:30.199145 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.215962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.216034 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.216058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.216089 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.216113 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.319751 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.319829 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.319845 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.319872 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.319887 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.423841 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.423905 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.423922 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.423948 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.423964 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.527058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.527110 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.527125 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.527147 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.527162 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.629776 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.629812 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.629819 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.629834 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.629844 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.732613 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.732658 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.732667 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.732682 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.732692 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.835469 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.835531 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.835548 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.835571 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.835589 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.938592 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.938649 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.938669 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.938694 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:30 crc kubenswrapper[4979]: I1204 11:44:30.938713 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:30Z","lastTransitionTime":"2025-12-04T11:44:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.042680 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.042729 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.042741 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.042761 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.042773 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.146838 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.146905 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.146918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.146942 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.146958 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.250583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.250642 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.250897 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.251021 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.251046 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.356295 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.356381 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.356396 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.356415 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.356432 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.459247 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.459288 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.459333 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.459364 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.459381 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.563178 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.563230 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.563251 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.563277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.563296 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.666271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.666388 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.666412 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.666446 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.666469 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.768978 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.769008 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.769018 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.769032 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.769043 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.872287 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.872360 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.872374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.872396 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.872747 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.976328 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.976404 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.976417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.976436 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:31 crc kubenswrapper[4979]: I1204 11:44:31.976446 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:31Z","lastTransitionTime":"2025-12-04T11:44:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.078868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.078926 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.078936 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.078955 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.078968 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.181570 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.181626 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.181638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.181654 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.181668 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.198429 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.198517 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:32 crc kubenswrapper[4979]: E1204 11:44:32.198586 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.198660 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.198802 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:32 crc kubenswrapper[4979]: E1204 11:44:32.198807 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:32 crc kubenswrapper[4979]: E1204 11:44:32.198887 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:32 crc kubenswrapper[4979]: E1204 11:44:32.198980 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.284905 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.284966 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.284977 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.284998 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.285009 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.387951 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.388000 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.388013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.388032 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.388044 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.490887 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.490937 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.490949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.490966 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.490977 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.593885 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.593920 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.593929 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.593943 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.593954 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.631182 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:32 crc kubenswrapper[4979]: E1204 11:44:32.631385 4979 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:44:32 crc kubenswrapper[4979]: E1204 11:44:32.631455 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs podName:e031b763-d441-46af-a2af-ad62bd130a6c nodeName:}" failed. No retries permitted until 2025-12-04 11:45:36.631435259 +0000 UTC m=+160.905731063 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs") pod "network-metrics-daemon-td9cq" (UID: "e031b763-d441-46af-a2af-ad62bd130a6c") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.697206 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.697258 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.697276 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.697328 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.697349 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.800618 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.800677 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.800691 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.800715 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.800728 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.903805 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.903877 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.903911 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.903943 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:32 crc kubenswrapper[4979]: I1204 11:44:32.903969 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:32Z","lastTransitionTime":"2025-12-04T11:44:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.014601 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.014717 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.014734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.014763 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.014777 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.118384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.118467 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.118486 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.118519 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.118540 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.221182 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.221230 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.221239 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.221260 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.221269 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.324922 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.324968 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.324978 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.324995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.325012 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.429614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.429651 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.429664 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.429685 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.429699 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.533679 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.533762 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.533777 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.533798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.533812 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.636821 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.636861 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.636869 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.636887 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.636898 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.740553 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.740638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.740657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.740686 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.740708 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.842885 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.842934 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.842945 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.842962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.842972 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.945453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.945511 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.945525 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.945546 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:33 crc kubenswrapper[4979]: I1204 11:44:33.945560 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:33Z","lastTransitionTime":"2025-12-04T11:44:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.049009 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.049061 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.049072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.049092 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.049105 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.153182 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.153236 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.153248 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.153271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.153286 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.198496 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.198563 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.198575 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.198747 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:34 crc kubenswrapper[4979]: E1204 11:44:34.198951 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:34 crc kubenswrapper[4979]: E1204 11:44:34.199444 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:34 crc kubenswrapper[4979]: E1204 11:44:34.199591 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:34 crc kubenswrapper[4979]: E1204 11:44:34.199667 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.199972 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:44:34 crc kubenswrapper[4979]: E1204 11:44:34.200221 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.256907 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.256964 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.256978 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.256995 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.257007 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.359732 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.360436 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.360567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.360644 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.360710 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.464357 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.464735 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.464811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.464894 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.464974 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.567384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.568281 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.568420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.568538 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.568632 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.671409 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.671720 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.671800 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.671906 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.672009 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.775013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.775069 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.775081 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.775104 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.775118 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.876918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.877367 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.877475 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.877587 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.877684 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.981226 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.981773 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.981874 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.981987 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:34 crc kubenswrapper[4979]: I1204 11:44:34.982109 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:34Z","lastTransitionTime":"2025-12-04T11:44:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.085777 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.085826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.085847 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.085867 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.085882 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.189739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.189793 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.189806 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.189826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.189837 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.293123 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.293188 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.293200 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.293222 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.293235 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.395800 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.395869 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.395883 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.395904 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.395917 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.498821 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.498881 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.498897 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.498921 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.498936 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.602355 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.602435 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.602453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.602480 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.602497 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.705413 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.705483 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.705505 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.705533 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.705555 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.808701 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.808765 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.808783 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.808813 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.808845 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.912794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.912868 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.912879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.912899 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:35 crc kubenswrapper[4979]: I1204 11:44:35.912911 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:35Z","lastTransitionTime":"2025-12-04T11:44:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.016191 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.016263 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.016277 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.016337 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.016353 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.119420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.119465 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.119476 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.119496 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.119510 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.199091 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.199192 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.199291 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.199413 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:36 crc kubenswrapper[4979]: E1204 11:44:36.199405 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:36 crc kubenswrapper[4979]: E1204 11:44:36.199502 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:36 crc kubenswrapper[4979]: E1204 11:44:36.199608 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:36 crc kubenswrapper[4979]: E1204 11:44:36.199799 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.222038 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.222113 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.222135 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.222161 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.222179 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.223981 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"39776bb9-499a-4b1f-9a0e-28d495281b94\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://962fbc29ebecb3f363684da33f6f2f5e5c12749f45243974977762ede0e898e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a8aca715f1acca4fcbf0759c160ef19e38165b6ee4269e42c4228b4a73bdfe1e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2668ae34e06d39d82c690a4056aefcf511fc157721bbcc4ecee5d5c670d202d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://026c6f5c83d85704894c6b067c7de8ec469c8983c8c91c5675044592f6013a69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://091235e59172b3679438926ca4d3dd032e406323def82332ac396efb5fb289ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9cf544efca8b14f7f1ee74c0cbf51a4d38cc68fe5802940a0d1422d61d2fd2d2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de9a7104265878473c3877c4a756a1396374bd7d36a654a0efbc4bf6d623fd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://80018c16664cbea61eae75bbbfaa25c2912e7224a72d33e6f1cbc7a46d76a65f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.246894 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:17Z\\\",\\\"message\\\":\\\"ba0} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:17Z is after 2025-08-24T17:21:41Z]\\\\nI1204 11:44:17.379359 7043 services_controller.go:473] Services do not match for network=default, existing lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"5b85277d-d9b7-4a68-8e4e-2b80594d9347\\\\\\\", Protocol:\\\\\\\"tcp\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-machine-config-operator/machine-config-operator\\\\\\\"}, Opts:services.LBOpts{Reject:false, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{}, Templates:services.TemplateMap{}, Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}, built lbs: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-machine-config-operator/machine-config-operator_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:44:16Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6h6d6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-x6k66\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.263411 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d9b48d97-5715-467e-b044-4812d6dfea63\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://94358ebeae4b076cab81dd23719e2af8cf29cad40f62924693d449b20052565a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8x6hm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-8twmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.277566 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a19f9554-b031-4d5c-a4aa-45f519d09945\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7201ad76a88f3e437416b55fbc2ffeceb6e438835722624bca8fff1d73fe739\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5186995f78a8588d1c726e0f4c00305a34822400fba156cec0b309c32e2a2394\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f575g\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:27Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-jf25c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.295183 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8378641c-2a6c-4f25-a047-f928917e0df5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"message\\\":\\\"le observer\\\\nW1204 11:43:13.232289 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1204 11:43:13.232478 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1204 11:43:13.233661 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2200873822/tls.crt::/tmp/serving-cert-2200873822/tls.key\\\\\\\"\\\\nI1204 11:43:13.681546 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1204 11:43:13.685793 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1204 11:43:13.685822 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1204 11:43:13.685865 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1204 11:43:13.685871 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1204 11:43:13.690932 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1204 11:43:13.690961 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690967 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1204 11:43:13.690972 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1204 11:43:13.690979 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1204 11:43:13.690986 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1204 11:43:13.690990 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1204 11:43:13.691157 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1204 11:43:13.693112 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:58Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.314894 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7d612fff-7871-4115-984c-c7e729b9f95f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c5d6f04570920d456f295fbfe54524e331ab365c006c28ffb8aca111ee3ae6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b3f965bf68ff7a3c61beec4b2955945c9ffed1868cc896fafa08e22e7a4e2d28\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a21c03a19327b94f31f2ff559c810e705225e1ffd62e52b03ee33c89be7b3d4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.325380 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.325455 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.325466 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.325485 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.325496 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.330690 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e4daf0bfaa8f3f4cae0564c0adb8e01137c116c97fafeb10c3e10074e999b506\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.347517 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.367537 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b7ecfd783e49a436271f211e4327f1bcee83c6340d18f875d38e614f06b190df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1bb937798d05fc3da817bfab7bfff3a75a743b3b8c973ec3fce9289534316745\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.391541 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-95gj9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c848f116-de32-47fa-9f47-5c0ff8204d29\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13c2d58c1a3e2b90974d3f4d7461dcddc432dd7bff4a8cbf1ebcea0482f7d872\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4bd5b0990ad1e18401d4f656ef55703a835fea0745601be866ddc4358e92f16c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00c37c2daffa875d8541c69a7c0ea93e0ad15815389402ec58c9c48682c163e5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://75e03075606b9f65687b329e6d3dbaf666bf6b1bd25bfcdc40d164daf67ccc80\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://20cab929e3ea49ac834ee98f939f5acf8340202ef93315d3bea76d744063f39f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78c9d2c89dd2a97e35a41615bfb157d7999514380f2e7541fc4a70cd68f85424\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://77c1f9956b19df91fcbd017c9e257662f108bdd8c83a5a790d49665688b62afd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:43:21Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9x7gq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-95gj9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.407721 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b87b21ad-7aaa-443d-b9e5-984c87d2d20f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f3981c91bd71a8dd3efbb1c823612de36e047acfc9f9cd6f76ed1904789d15c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2dc08289eddf3b819e1ab655e3cfd253a41599061c8153c294752683e0c8e1c4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.427093 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e9233ec0-115e-4f1a-8e40-ba57f5af1e18\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:42:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7b6a7b67f35c441fbc6c8b1adaa491414be33032cdf9a450a9538a7471da7a46\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7cee30d13710861bea84fe0e5e00a7f74030de55c24e31e5479ef55932821438\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7196ce0a15d776e7319086b1d7e7bdf11d82cebba9be3f4d3924fe68fcbb847\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://203c7db6626bda63df81983a132b6e3bdcd74c0170752ef7f93539b0981c71b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-04T11:42:57Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-04T11:42:57Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:42:56Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.430444 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.430706 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.430869 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.431003 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.431168 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.442551 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.460099 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:18Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9f454d8489a525065519c1ef256c6657fc9cbe9151c8bf22ab4dd563b4061f0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.473259 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-smkn7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0dd48cc7-5561-4c17-9f53-f1e0c30129c2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://02762bc7c0e1f8cf49874b84b39c50b087f556d842e6a4c1d497add13af7748d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bmq4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-smkn7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.488211 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-g4nk7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"12339390-77cf-4906-80c3-ddb729e42132\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-04T11:44:00Z\\\",\\\"message\\\":\\\"2025-12-04T11:43:15+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c\\\\n2025-12-04T11:43:15+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c9a307f9-1056-4497-bc3c-a2d7483bb09c to /host/opt/cni/bin/\\\\n2025-12-04T11:43:15Z [verbose] multus-daemon started\\\\n2025-12-04T11:43:15Z [verbose] Readiness Indicator file check\\\\n2025-12-04T11:44:00Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-04T11:43:15Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:44:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9p5sn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-g4nk7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.505161 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-td9cq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e031b763-d441-46af-a2af-ad62bd130a6c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:28Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsp9b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:28Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-td9cq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.521632 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.533470 4979 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-x7qhm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c168d89e-7737-464e-a932-41993ad0c931\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-04T11:43:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14a83587854fd139c489da2e50cc66e3a48f3ed3be87f7926370d4ea787f0e60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-04T11:43:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ntxjs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-04T11:43:16Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-x7qhm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:36Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.534772 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.534817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.534832 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.534856 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.534871 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.637643 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.637680 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.637688 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.637701 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.637713 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.740040 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.740107 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.740121 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.740144 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.740158 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.843720 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.843797 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.843821 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.843853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.843878 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.947425 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.947501 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.947512 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.947533 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:36 crc kubenswrapper[4979]: I1204 11:44:36.947548 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:36Z","lastTransitionTime":"2025-12-04T11:44:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.050681 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.050734 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.050750 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.050774 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.050791 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.153959 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.154027 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.154042 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.154063 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.154078 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.256921 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.256973 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.256986 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.257006 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.257019 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.360922 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.360986 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.361002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.361021 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.361034 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.463987 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.464042 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.464061 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.464084 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.464100 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.566974 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.567032 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.567049 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.567075 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.567094 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.671238 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.671284 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.671321 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.671342 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.671357 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.775424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.776010 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.776408 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.776648 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.776912 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.880580 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.880646 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.880666 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.880692 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.880711 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.982737 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.983179 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.983265 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.983364 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:37 crc kubenswrapper[4979]: I1204 11:44:37.983442 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:37Z","lastTransitionTime":"2025-12-04T11:44:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.086953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.087448 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.087551 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.087579 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.087608 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.190224 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.190287 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.190319 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.190337 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.190350 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.198752 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.198880 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.198763 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.198962 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.198875 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.199124 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.199442 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.199529 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.293530 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.293620 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.293630 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.293644 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.293654 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.395548 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.395582 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.395590 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.395604 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.395615 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.498164 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.498492 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.498638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.498842 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.499042 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.600929 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.600961 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.600971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.600984 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.600994 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.682486 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.682573 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.682593 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.682625 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.682647 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.701375 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:38Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.709081 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.709150 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.709164 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.709185 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.709198 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.729632 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:38Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.734250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.734361 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.734374 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.734393 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.734432 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.751084 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:38Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.756636 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.756932 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.757105 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.757384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.757569 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.775996 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:38Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.780925 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.780988 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.781005 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.781028 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.781045 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.800071 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-04T11:44:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"7408354e-3656-48ab-94ea-2745a060a897\\\",\\\"systemUUID\\\":\\\"eae98eec-36de-419b-b26f-0e75001a46a0\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-04T11:44:38Z is after 2025-08-24T17:21:41Z" Dec 04 11:44:38 crc kubenswrapper[4979]: E1204 11:44:38.800402 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.802897 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.802952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.802963 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.802979 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.802991 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.904849 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.904905 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.904914 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.904930 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:38 crc kubenswrapper[4979]: I1204 11:44:38.904941 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:38Z","lastTransitionTime":"2025-12-04T11:44:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.008010 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.008068 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.008081 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.008103 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.008114 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.111917 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.111966 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.111981 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.112005 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.112021 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.215279 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.215372 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.215389 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.215413 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.215432 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.318367 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.318416 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.318427 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.318445 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.318458 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.787962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.788014 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.788029 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.788048 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.788061 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.891046 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.891453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.891777 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.892058 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.892285 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.995079 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.995551 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.995763 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.995994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:39 crc kubenswrapper[4979]: I1204 11:44:39.996199 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:39Z","lastTransitionTime":"2025-12-04T11:44:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.099120 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.099563 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.100103 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.100285 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.100451 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.198760 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.198825 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.198658 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.198790 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:40 crc kubenswrapper[4979]: E1204 11:44:40.198958 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:40 crc kubenswrapper[4979]: E1204 11:44:40.199076 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:40 crc kubenswrapper[4979]: E1204 11:44:40.199161 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:40 crc kubenswrapper[4979]: E1204 11:44:40.199351 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.204280 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.204381 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.204407 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.204437 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.204596 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.307983 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.308524 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.308542 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.308567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.308587 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.411583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.411629 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.411641 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.411661 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.411677 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.514907 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.515465 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.515568 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.515656 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.515747 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.620460 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.620530 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.620543 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.620616 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.620633 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.723129 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.723572 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.723676 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.723774 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.723840 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.826692 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.827370 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.827387 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.827404 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.827416 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.931229 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.931271 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.931280 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.931337 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:40 crc kubenswrapper[4979]: I1204 11:44:40.931350 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:40Z","lastTransitionTime":"2025-12-04T11:44:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.041325 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.041383 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.041393 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.041410 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.041421 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.144395 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.144464 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.144481 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.144507 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.144527 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.247759 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.247821 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.247835 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.247854 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.247869 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.352350 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.352393 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.352404 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.352428 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.352442 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.455356 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.455417 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.455433 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.455458 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.455472 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.558230 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.558359 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.558384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.558420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.558445 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.661126 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.661195 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.661216 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.661241 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.661256 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.764910 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.764953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.764962 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.764983 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.765002 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.868529 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.868589 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.868605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.868629 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.868646 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.971818 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.972222 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.972336 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.972424 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:41 crc kubenswrapper[4979]: I1204 11:44:41.972510 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:41Z","lastTransitionTime":"2025-12-04T11:44:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.075420 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.075798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.075908 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.076030 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.076122 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.180189 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.180250 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.180266 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.180326 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.180347 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.198851 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.198889 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:42 crc kubenswrapper[4979]: E1204 11:44:42.199137 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.198912 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:42 crc kubenswrapper[4979]: E1204 11:44:42.199465 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:42 crc kubenswrapper[4979]: E1204 11:44:42.199606 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.199509 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:42 crc kubenswrapper[4979]: E1204 11:44:42.199818 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.282742 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.283209 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.283415 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.283605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.283794 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.388728 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.389442 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.389484 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.389510 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.389534 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.492459 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.492506 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.492516 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.492533 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.492544 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.595716 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.596135 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.596358 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.596598 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.596927 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.699940 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.700378 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.700503 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.700590 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.700707 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.803716 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.803768 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.803781 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.803798 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.803810 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.907263 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.907339 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.907356 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.907378 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:42 crc kubenswrapper[4979]: I1204 11:44:42.907392 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:42Z","lastTransitionTime":"2025-12-04T11:44:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.010521 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.010578 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.010608 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.010657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.010677 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.114171 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.114244 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.114265 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.114295 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.114362 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.218691 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.218744 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.218765 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.218785 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.218800 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.321719 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.321803 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.321828 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.321858 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.321879 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.424994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.425073 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.425091 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.425118 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.425138 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.528061 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.528168 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.528192 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.528224 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.528250 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.631766 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.631825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.631836 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.631853 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.631864 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.734800 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.734876 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.734903 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.734933 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.734956 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.840175 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.840269 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.840328 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.840368 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.840406 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.943761 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.944874 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.945181 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.946542 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:43 crc kubenswrapper[4979]: I1204 11:44:43.946618 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:43Z","lastTransitionTime":"2025-12-04T11:44:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.050550 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.050617 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.050634 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.050660 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.050677 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.154628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.154701 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.154718 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.154741 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.154755 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.198595 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.198705 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.198741 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:44 crc kubenswrapper[4979]: E1204 11:44:44.198876 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.198959 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:44 crc kubenswrapper[4979]: E1204 11:44:44.199067 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:44 crc kubenswrapper[4979]: E1204 11:44:44.199172 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:44 crc kubenswrapper[4979]: E1204 11:44:44.199354 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.258717 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.258782 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.258794 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.258818 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.258833 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.362188 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.362268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.362333 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.362387 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.362433 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.465882 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.465957 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.465971 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.465997 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.466013 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.569284 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.569352 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.569365 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.569382 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.569393 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.672245 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.672344 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.672359 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.672384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.672402 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.775755 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.775805 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.775817 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.775836 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.775851 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.878823 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.878885 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.878899 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.878920 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.878934 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.982455 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.982494 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.982504 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.982521 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:44 crc kubenswrapper[4979]: I1204 11:44:44.982536 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:44Z","lastTransitionTime":"2025-12-04T11:44:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.085779 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.085826 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.085837 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.086051 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.086061 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.188924 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.188993 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.189006 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.189033 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.189050 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.291489 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.291555 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.291567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.291591 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.291606 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.394268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.394400 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.394419 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.394477 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.394497 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.498354 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.498392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.498403 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.498419 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.498430 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.600810 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.600859 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.600873 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.600896 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.600909 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.704260 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.704326 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.704337 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.704356 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.704369 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.807887 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.807949 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.807979 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.808028 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.808057 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.911384 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.911464 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.911491 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.911595 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:45 crc kubenswrapper[4979]: I1204 11:44:45.911620 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:45Z","lastTransitionTime":"2025-12-04T11:44:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.015146 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.015203 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.015215 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.015235 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.015252 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.118170 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.118232 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.118247 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.118267 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.118280 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.198727 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.198782 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.198727 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:46 crc kubenswrapper[4979]: E1204 11:44:46.199093 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.199200 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:46 crc kubenswrapper[4979]: E1204 11:44:46.200331 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:46 crc kubenswrapper[4979]: E1204 11:44:46.199540 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:46 crc kubenswrapper[4979]: E1204 11:44:46.199361 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.222572 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.222613 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.222623 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.222638 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.222648 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.243204 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=62.243183273 podStartE2EDuration="1m2.243183273s" podCreationTimestamp="2025-12-04 11:43:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.225801977 +0000 UTC m=+110.500097781" watchObservedRunningTime="2025-12-04 11:44:46.243183273 +0000 UTC m=+110.517479077" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.290253 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-smkn7" podStartSLOduration=93.290227985 podStartE2EDuration="1m33.290227985s" podCreationTimestamp="2025-12-04 11:43:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.272618864 +0000 UTC m=+110.546914688" watchObservedRunningTime="2025-12-04 11:44:46.290227985 +0000 UTC m=+110.564523789" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.303653 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-g4nk7" podStartSLOduration=92.303628847 podStartE2EDuration="1m32.303628847s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.290663127 +0000 UTC m=+110.564958931" watchObservedRunningTime="2025-12-04 11:44:46.303628847 +0000 UTC m=+110.577924651" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.325334 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.325382 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.325392 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.325406 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.325418 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.331691 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.331675302 podStartE2EDuration="31.331675302s" podCreationTimestamp="2025-12-04 11:44:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.318217869 +0000 UTC m=+110.592513673" watchObservedRunningTime="2025-12-04 11:44:46.331675302 +0000 UTC m=+110.605971106" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.345214 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-x7qhm" podStartSLOduration=92.345187566 podStartE2EDuration="1m32.345187566s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.331930708 +0000 UTC m=+110.606226502" watchObservedRunningTime="2025-12-04 11:44:46.345187566 +0000 UTC m=+110.619483370" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.403195 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podStartSLOduration=92.403164875 podStartE2EDuration="1m32.403164875s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.389153148 +0000 UTC m=+110.663448952" watchObservedRunningTime="2025-12-04 11:44:46.403164875 +0000 UTC m=+110.677460679" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.427764 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.427809 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.427822 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.427838 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.427850 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.444097 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-jf25c" podStartSLOduration=91.444070447 podStartE2EDuration="1m31.444070447s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.404086229 +0000 UTC m=+110.678382033" watchObservedRunningTime="2025-12-04 11:44:46.444070447 +0000 UTC m=+110.718366251" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.444573 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=90.444564659 podStartE2EDuration="1m30.444564659s" podCreationTimestamp="2025-12-04 11:43:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.441663203 +0000 UTC m=+110.715959017" watchObservedRunningTime="2025-12-04 11:44:46.444564659 +0000 UTC m=+110.718860463" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.457744 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=93.457717734 podStartE2EDuration="1m33.457717734s" podCreationTimestamp="2025-12-04 11:43:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.457604171 +0000 UTC m=+110.731899995" watchObservedRunningTime="2025-12-04 11:44:46.457717734 +0000 UTC m=+110.732013538" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.525903 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-95gj9" podStartSLOduration=92.52585327 podStartE2EDuration="1m32.52585327s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.525521181 +0000 UTC m=+110.799817005" watchObservedRunningTime="2025-12-04 11:44:46.52585327 +0000 UTC m=+110.800149084" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.529932 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.530236 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.530327 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.530453 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.530525 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.546213 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=92.546192642 podStartE2EDuration="1m32.546192642s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:46.546150101 +0000 UTC m=+110.820445905" watchObservedRunningTime="2025-12-04 11:44:46.546192642 +0000 UTC m=+110.820488436" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.633994 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.634056 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.634072 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.634097 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.634112 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.737031 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.737080 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.737091 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.737108 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.737119 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.841445 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.841499 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.841517 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.841551 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.841566 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.944723 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.944791 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.944805 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.944825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:46 crc kubenswrapper[4979]: I1204 11:44:46.944838 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:46Z","lastTransitionTime":"2025-12-04T11:44:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.047244 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.047290 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.047312 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.047331 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.047342 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.151212 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.151269 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.151283 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.151320 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.151340 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.198775 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:44:47 crc kubenswrapper[4979]: E1204 11:44:47.198977 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-x6k66_openshift-ovn-kubernetes(af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26)\"" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.255204 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.255251 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.255261 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.255281 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.255292 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.358681 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.358729 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.358744 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.358765 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.358777 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.461524 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.461573 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.461586 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.461605 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.461619 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.564759 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.564811 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.564901 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.564927 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.564943 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.668744 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.668787 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.668801 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.668822 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.668835 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.772057 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.772221 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.772241 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.772628 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.772927 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.876089 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.876153 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.876167 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.876187 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.876199 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.933823 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/1.log" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.934650 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/0.log" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.934749 4979 generic.go:334] "Generic (PLEG): container finished" podID="12339390-77cf-4906-80c3-ddb729e42132" containerID="40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93" exitCode=1 Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.934806 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerDied","Data":"40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93"} Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.934874 4979 scope.go:117] "RemoveContainer" containerID="dcf8cec90fb2c2a2eddcea54c9bc6194f615e623d66ecc3e3a55df82b0f3a1a6" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.935404 4979 scope.go:117] "RemoveContainer" containerID="40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93" Dec 04 11:44:47 crc kubenswrapper[4979]: E1204 11:44:47.935630 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-g4nk7_openshift-multus(12339390-77cf-4906-80c3-ddb729e42132)\"" pod="openshift-multus/multus-g4nk7" podUID="12339390-77cf-4906-80c3-ddb729e42132" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.979953 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.980002 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.980013 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.980033 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:47 crc kubenswrapper[4979]: I1204 11:44:47.980044 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:47Z","lastTransitionTime":"2025-12-04T11:44:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.083521 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.083583 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.083598 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.083622 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.083644 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.186790 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.187122 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.187198 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.187268 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.187392 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.198265 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:48 crc kubenswrapper[4979]: E1204 11:44:48.198528 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.198362 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:48 crc kubenswrapper[4979]: E1204 11:44:48.198728 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.198265 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:48 crc kubenswrapper[4979]: E1204 11:44:48.199095 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.198362 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:48 crc kubenswrapper[4979]: E1204 11:44:48.199380 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.290078 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.290129 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.290138 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.290160 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.290174 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.392657 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.392714 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.392726 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.392745 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.392759 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.495739 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.495879 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.495918 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.495952 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.495977 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.600397 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.600483 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.600511 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.600561 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.600586 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.704100 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.704621 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.704825 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.705022 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.705198 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.808414 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.808492 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.808560 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.808614 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.808641 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.902655 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.902708 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.902727 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.902750 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.902769 4979 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-04T11:44:48Z","lastTransitionTime":"2025-12-04T11:44:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.942197 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/1.log" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.959710 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b"] Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.964463 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.967373 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.968040 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.969555 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.970042 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.992050 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/039212f2-c983-4955-b50c-7ea9efe235ee-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.992097 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039212f2-c983-4955-b50c-7ea9efe235ee-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.992146 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/039212f2-c983-4955-b50c-7ea9efe235ee-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.992164 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/039212f2-c983-4955-b50c-7ea9efe235ee-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:48 crc kubenswrapper[4979]: I1204 11:44:48.992198 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/039212f2-c983-4955-b50c-7ea9efe235ee-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093705 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039212f2-c983-4955-b50c-7ea9efe235ee-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093783 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/039212f2-c983-4955-b50c-7ea9efe235ee-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093801 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/039212f2-c983-4955-b50c-7ea9efe235ee-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093854 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/039212f2-c983-4955-b50c-7ea9efe235ee-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093895 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/039212f2-c983-4955-b50c-7ea9efe235ee-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093973 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/039212f2-c983-4955-b50c-7ea9efe235ee-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.093998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/039212f2-c983-4955-b50c-7ea9efe235ee-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.095001 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/039212f2-c983-4955-b50c-7ea9efe235ee-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.102238 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/039212f2-c983-4955-b50c-7ea9efe235ee-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.116100 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/039212f2-c983-4955-b50c-7ea9efe235ee-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8g42b\" (UID: \"039212f2-c983-4955-b50c-7ea9efe235ee\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.279632 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" Dec 04 11:44:49 crc kubenswrapper[4979]: W1204 11:44:49.315115 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod039212f2_c983_4955_b50c_7ea9efe235ee.slice/crio-395b02194a5ee86be982748948ee41bea58ba017bdc0eab7dc6d81efc5e79738 WatchSource:0}: Error finding container 395b02194a5ee86be982748948ee41bea58ba017bdc0eab7dc6d81efc5e79738: Status 404 returned error can't find the container with id 395b02194a5ee86be982748948ee41bea58ba017bdc0eab7dc6d81efc5e79738 Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.947287 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" event={"ID":"039212f2-c983-4955-b50c-7ea9efe235ee","Type":"ContainerStarted","Data":"ef0c52497cefbe03558387713acac687296100387fb212bc95e3187afc7317c1"} Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.947356 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" event={"ID":"039212f2-c983-4955-b50c-7ea9efe235ee","Type":"ContainerStarted","Data":"395b02194a5ee86be982748948ee41bea58ba017bdc0eab7dc6d81efc5e79738"} Dec 04 11:44:49 crc kubenswrapper[4979]: I1204 11:44:49.965263 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8g42b" podStartSLOduration=95.964750728 podStartE2EDuration="1m35.964750728s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:49.964022839 +0000 UTC m=+114.238318663" watchObservedRunningTime="2025-12-04 11:44:49.964750728 +0000 UTC m=+114.239046532" Dec 04 11:44:50 crc kubenswrapper[4979]: I1204 11:44:50.198820 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:50 crc kubenswrapper[4979]: E1204 11:44:50.198970 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:50 crc kubenswrapper[4979]: I1204 11:44:50.199189 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:50 crc kubenswrapper[4979]: E1204 11:44:50.199238 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:50 crc kubenswrapper[4979]: I1204 11:44:50.199369 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:50 crc kubenswrapper[4979]: E1204 11:44:50.199415 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:50 crc kubenswrapper[4979]: I1204 11:44:50.199882 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:50 crc kubenswrapper[4979]: E1204 11:44:50.199948 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:52 crc kubenswrapper[4979]: I1204 11:44:52.198606 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:52 crc kubenswrapper[4979]: I1204 11:44:52.198651 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:52 crc kubenswrapper[4979]: I1204 11:44:52.198606 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:52 crc kubenswrapper[4979]: E1204 11:44:52.198761 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:52 crc kubenswrapper[4979]: I1204 11:44:52.199011 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:52 crc kubenswrapper[4979]: E1204 11:44:52.198984 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:52 crc kubenswrapper[4979]: E1204 11:44:52.199070 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:52 crc kubenswrapper[4979]: E1204 11:44:52.199134 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:54 crc kubenswrapper[4979]: I1204 11:44:54.198522 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:54 crc kubenswrapper[4979]: E1204 11:44:54.198684 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:54 crc kubenswrapper[4979]: I1204 11:44:54.198942 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:54 crc kubenswrapper[4979]: E1204 11:44:54.199007 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:54 crc kubenswrapper[4979]: I1204 11:44:54.199174 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:54 crc kubenswrapper[4979]: E1204 11:44:54.199240 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:54 crc kubenswrapper[4979]: I1204 11:44:54.199395 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:54 crc kubenswrapper[4979]: E1204 11:44:54.199529 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:56 crc kubenswrapper[4979]: I1204 11:44:56.198865 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:56 crc kubenswrapper[4979]: I1204 11:44:56.198859 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:56 crc kubenswrapper[4979]: I1204 11:44:56.198889 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:56 crc kubenswrapper[4979]: I1204 11:44:56.198940 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:56 crc kubenswrapper[4979]: E1204 11:44:56.200155 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:56 crc kubenswrapper[4979]: E1204 11:44:56.200263 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:56 crc kubenswrapper[4979]: E1204 11:44:56.200396 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:56 crc kubenswrapper[4979]: E1204 11:44:56.200559 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:56 crc kubenswrapper[4979]: E1204 11:44:56.244485 4979 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 04 11:44:56 crc kubenswrapper[4979]: E1204 11:44:56.302117 4979 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.199570 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.199701 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.199727 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.199727 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:58 crc kubenswrapper[4979]: E1204 11:44:58.199867 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:44:58 crc kubenswrapper[4979]: E1204 11:44:58.200029 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.200177 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:44:58 crc kubenswrapper[4979]: E1204 11:44:58.200186 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:44:58 crc kubenswrapper[4979]: E1204 11:44:58.200343 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.988813 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/3.log" Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.992097 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerStarted","Data":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} Dec 04 11:44:58 crc kubenswrapper[4979]: I1204 11:44:58.993104 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:44:59 crc kubenswrapper[4979]: I1204 11:44:59.005920 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-td9cq"] Dec 04 11:44:59 crc kubenswrapper[4979]: I1204 11:44:59.006073 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:44:59 crc kubenswrapper[4979]: E1204 11:44:59.006289 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:44:59 crc kubenswrapper[4979]: I1204 11:44:59.028339 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podStartSLOduration=105.028314457 podStartE2EDuration="1m45.028314457s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:44:59.02648865 +0000 UTC m=+123.300784474" watchObservedRunningTime="2025-12-04 11:44:59.028314457 +0000 UTC m=+123.302610261" Dec 04 11:44:59 crc kubenswrapper[4979]: I1204 11:44:59.198865 4979 scope.go:117] "RemoveContainer" containerID="40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93" Dec 04 11:44:59 crc kubenswrapper[4979]: I1204 11:44:59.996705 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/1.log" Dec 04 11:44:59 crc kubenswrapper[4979]: I1204 11:44:59.997229 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerStarted","Data":"9a2f45253ea82811c08648bce91940b55482b206052274ef16b1d240ef1dda0c"} Dec 04 11:45:00 crc kubenswrapper[4979]: I1204 11:45:00.199483 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:00 crc kubenswrapper[4979]: I1204 11:45:00.199502 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:00 crc kubenswrapper[4979]: E1204 11:45:00.199659 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:45:00 crc kubenswrapper[4979]: I1204 11:45:00.199764 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:00 crc kubenswrapper[4979]: E1204 11:45:00.199919 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:45:00 crc kubenswrapper[4979]: I1204 11:45:00.200029 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:00 crc kubenswrapper[4979]: E1204 11:45:00.200420 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:45:00 crc kubenswrapper[4979]: E1204 11:45:00.200514 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:45:01 crc kubenswrapper[4979]: E1204 11:45:01.303620 4979 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 11:45:02 crc kubenswrapper[4979]: I1204 11:45:02.199066 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:02 crc kubenswrapper[4979]: I1204 11:45:02.199157 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:02 crc kubenswrapper[4979]: E1204 11:45:02.199262 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:45:02 crc kubenswrapper[4979]: E1204 11:45:02.199441 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:45:02 crc kubenswrapper[4979]: I1204 11:45:02.199075 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:02 crc kubenswrapper[4979]: E1204 11:45:02.199588 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:45:02 crc kubenswrapper[4979]: I1204 11:45:02.199758 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:02 crc kubenswrapper[4979]: E1204 11:45:02.199932 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:45:04 crc kubenswrapper[4979]: I1204 11:45:04.198933 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:04 crc kubenswrapper[4979]: I1204 11:45:04.198996 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:04 crc kubenswrapper[4979]: I1204 11:45:04.199051 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:04 crc kubenswrapper[4979]: E1204 11:45:04.199092 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:45:04 crc kubenswrapper[4979]: I1204 11:45:04.199246 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:04 crc kubenswrapper[4979]: E1204 11:45:04.199244 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:45:04 crc kubenswrapper[4979]: E1204 11:45:04.199375 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:45:04 crc kubenswrapper[4979]: E1204 11:45:04.199510 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:45:06 crc kubenswrapper[4979]: I1204 11:45:06.198998 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:06 crc kubenswrapper[4979]: I1204 11:45:06.199043 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:06 crc kubenswrapper[4979]: I1204 11:45:06.199043 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:06 crc kubenswrapper[4979]: E1204 11:45:06.200525 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 04 11:45:06 crc kubenswrapper[4979]: I1204 11:45:06.200755 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:06 crc kubenswrapper[4979]: E1204 11:45:06.200907 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 04 11:45:06 crc kubenswrapper[4979]: E1204 11:45:06.200975 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 04 11:45:06 crc kubenswrapper[4979]: E1204 11:45:06.200941 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-td9cq" podUID="e031b763-d441-46af-a2af-ad62bd130a6c" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.198957 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.198971 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.199017 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.199144 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.202686 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.203062 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.203733 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.204009 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.204054 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 11:45:08 crc kubenswrapper[4979]: I1204 11:45:08.204352 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.181567 4979 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.231666 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhnlh"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.232230 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.234473 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dkvxc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.236608 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.239330 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.240104 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.242495 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4v569"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.244731 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.244781 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.244871 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fkw88"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.245619 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.245770 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.248383 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.260993 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.262124 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.266716 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.267004 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.267207 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.267406 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.270438 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.270534 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xmjpc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.270919 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271030 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.270969 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271079 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271158 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271188 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271505 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271537 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271629 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271669 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.271804 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.275937 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.278208 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.278648 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.282664 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.282905 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.286387 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.286415 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.286415 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.286857 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.287192 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5bsgf"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.287519 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.288328 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.288341 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.288392 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.290653 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.291365 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.292660 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.310978 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-h6d8x"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.311500 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vbkql"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.315517 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fnchd"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.316542 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.317460 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.317930 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.311501 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.311549 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.314751 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.317382 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.317887 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319082 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319135 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.323548 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319144 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319195 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.324024 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.324526 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.324879 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.325202 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.325368 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.325504 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.325652 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.325791 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.325928 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.328659 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-5wfhh"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.329799 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.330135 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.330425 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319200 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.330732 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319207 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.331031 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.331229 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319291 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319658 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319803 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319814 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.319901 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.320149 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.320152 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.320238 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.333411 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.333458 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.333800 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.320290 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.320534 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.320881 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.333220 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.335606 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-qwjms"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.335827 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.337400 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.357884 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.358575 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.358630 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359703 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-config\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359733 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359763 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-config\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359787 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-etcd-client\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359803 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-client-ca\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359850 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359871 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359902 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-serving-cert\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359921 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-config\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359938 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/74542077-889e-48b4-9af3-3f0ffde94fc8-node-pullsecrets\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359955 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-audit\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.359984 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-encryption-config\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74542077-889e-48b4-9af3-3f0ffde94fc8-audit-dir\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360020 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-client-ca\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360040 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360059 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-serving-cert\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360082 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-config\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360103 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv4n9\" (UniqueName: \"kubernetes.io/projected/6b7bb127-9196-4baa-b581-5ec01aa6da33-kube-api-access-tv4n9\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360120 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-audit-policies\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360172 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-etcd-serving-ca\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360191 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360211 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr6kb\" (UniqueName: \"kubernetes.io/projected/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-kube-api-access-kr6kb\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360232 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360256 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360273 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-424kk\" (UniqueName: \"kubernetes.io/projected/e31c1aff-c394-476a-888e-c0e877f59717-kube-api-access-424kk\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360295 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-service-ca-bundle\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360332 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vkpc\" (UniqueName: \"kubernetes.io/projected/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-kube-api-access-6vkpc\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360351 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360405 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb6l5\" (UniqueName: \"kubernetes.io/projected/74542077-889e-48b4-9af3-3f0ffde94fc8-kube-api-access-xb6l5\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360422 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b7bb127-9196-4baa-b581-5ec01aa6da33-serving-cert\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360439 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e31c1aff-c394-476a-888e-c0e877f59717-audit-dir\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360458 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360484 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-serving-cert\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360503 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360522 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360541 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360582 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.360601 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-image-import-ca\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.363044 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.363417 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.363770 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364156 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364245 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364479 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364666 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364797 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364824 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.364915 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.365085 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.365245 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.365345 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.368958 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.369826 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.370266 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.370816 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.371589 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.372229 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.372482 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.372648 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.372736 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.373264 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.373344 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.375862 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.376464 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pwx2p"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.376858 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.377097 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.377885 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.378217 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.379710 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.380140 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.380527 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.380681 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.380784 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.394081 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.394591 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.394633 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.394879 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.395250 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.395403 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.395590 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.395745 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.399360 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.399583 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.399738 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.399855 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.399937 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.400173 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.410842 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.412142 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-j84bh"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.412983 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.413200 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.413608 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.415267 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.415786 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.416977 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.421023 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.422253 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.423965 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.425911 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.428259 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.430071 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.435099 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.435196 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csp56"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.435895 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.436219 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.437125 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.439161 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-stbh4"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.439636 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.440954 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.441550 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.443121 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dgmfk"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.444203 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.444391 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.446239 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-67g6s"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.446717 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.446830 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.447063 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.448415 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhnlh"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.448444 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4v569"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.448517 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.453124 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xmjpc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.453173 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dkvxc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.458029 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.459240 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-h6d8x"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.463567 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.463707 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4gvd\" (UniqueName: \"kubernetes.io/projected/ee371f5c-b62b-4694-808b-f1f26a2353c5-kube-api-access-m4gvd\") pod \"downloads-7954f5f757-qwjms\" (UID: \"ee371f5c-b62b-4694-808b-f1f26a2353c5\") " pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.463740 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.463785 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-service-ca\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.463933 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3bd54a5-561b-4207-9c2b-24718a0b473f-config\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.463961 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8br8k\" (UniqueName: \"kubernetes.io/projected/5a9c53b8-81fc-4e05-bb9b-f205aea24e47-kube-api-access-8br8k\") pod \"dns-operator-744455d44c-h6d8x\" (UID: \"5a9c53b8-81fc-4e05-bb9b-f205aea24e47\") " pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464031 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-serving-cert\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464148 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fnchd"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464743 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-config\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464802 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1b73cd49-492e-443e-b199-b7f9b80b066c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c54gv\" (UID: \"1b73cd49-492e-443e-b199-b7f9b80b066c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464828 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-audit\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464959 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-etcd-client\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.464976 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2chqc\" (UniqueName: \"kubernetes.io/projected/5592da75-33af-4003-96eb-542cf6cdc256-kube-api-access-2chqc\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465099 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb732274-6b17-47f6-ace6-485706ecd538-serving-cert\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465132 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/74542077-889e-48b4-9af3-3f0ffde94fc8-node-pullsecrets\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465150 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-serving-cert\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465285 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5592da75-33af-4003-96eb-542cf6cdc256-audit-dir\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465322 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/74542077-889e-48b4-9af3-3f0ffde94fc8-node-pullsecrets\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465409 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-client\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465433 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-oauth-serving-cert\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465456 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-encryption-config\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465577 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74542077-889e-48b4-9af3-3f0ffde94fc8-audit-dir\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465737 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/74542077-889e-48b4-9af3-3f0ffde94fc8-audit-dir\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465764 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-client-ca\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465790 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465866 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465909 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-serving-cert\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465932 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-metrics-tls\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.465959 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbsr5\" (UniqueName: \"kubernetes.io/projected/82926c75-1f3c-4230-b981-f341f4e17ee2-kube-api-access-bbsr5\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466002 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c2fg\" (UniqueName: \"kubernetes.io/projected/45b72490-12d9-450f-9a00-2d467c68a35b-kube-api-access-5c2fg\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466021 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgblx\" (UniqueName: \"kubernetes.io/projected/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-kube-api-access-hgblx\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466039 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv4n9\" (UniqueName: \"kubernetes.io/projected/6b7bb127-9196-4baa-b581-5ec01aa6da33-kube-api-access-tv4n9\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466074 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-audit-policies\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466091 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-audit-policies\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466109 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4a8680c4-28b9-4242-8249-ad3efcd042a0-machine-approver-tls\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-config\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466173 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-ca\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466238 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45b72490-12d9-450f-9a00-2d467c68a35b-serving-cert\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466268 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-oauth-config\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466385 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8680c4-28b9-4242-8249-ad3efcd042a0-config\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466426 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466909 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9f27cd34-9226-4968-a4e7-83788b460159-images\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466932 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cct57\" (UniqueName: \"kubernetes.io/projected/1b73cd49-492e-443e-b199-b7f9b80b066c-kube-api-access-cct57\") pod \"cluster-samples-operator-665b6dd947-c54gv\" (UID: \"1b73cd49-492e-443e-b199-b7f9b80b066c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.466996 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-etcd-serving-ca\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467022 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467145 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4a8680c4-28b9-4242-8249-ad3efcd042a0-auth-proxy-config\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467166 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-trusted-ca-bundle\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467289 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr6kb\" (UniqueName: \"kubernetes.io/projected/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-kube-api-access-kr6kb\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467478 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467528 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-client-ca\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467529 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/749499f8-a6c6-403c-9a4a-26242d44b3c5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467581 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-service-ca\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467684 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-audit-policies\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467787 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-etcd-serving-ca\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467878 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467912 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-424kk\" (UniqueName: \"kubernetes.io/projected/e31c1aff-c394-476a-888e-c0e877f59717-kube-api-access-424kk\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.467939 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-encryption-config\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468211 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-audit\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468728 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vkpc\" (UniqueName: \"kubernetes.io/projected/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-kube-api-access-6vkpc\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468768 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-config\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468819 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f27cd34-9226-4968-a4e7-83788b460159-config\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468836 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468884 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-config\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.468921 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/749499f8-a6c6-403c-9a4a-26242d44b3c5-config\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469021 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3bd54a5-561b-4207-9c2b-24718a0b473f-serving-cert\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469043 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469061 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-service-ca-bundle\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469209 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/749499f8-a6c6-403c-9a4a-26242d44b3c5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469249 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-serving-cert\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469309 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-config\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469333 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvqkn\" (UniqueName: \"kubernetes.io/projected/4a8680c4-28b9-4242-8249-ad3efcd042a0-kube-api-access-lvqkn\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469360 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b7bb127-9196-4baa-b581-5ec01aa6da33-serving-cert\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469380 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e31c1aff-c394-476a-888e-c0e877f59717-audit-dir\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469416 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e31c1aff-c394-476a-888e-c0e877f59717-audit-dir\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469449 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469532 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3bd54a5-561b-4207-9c2b-24718a0b473f-trusted-ca\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469653 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb6l5\" (UniqueName: \"kubernetes.io/projected/74542077-889e-48b4-9af3-3f0ffde94fc8-kube-api-access-xb6l5\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469688 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-trusted-ca\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469842 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzdhh\" (UniqueName: \"kubernetes.io/projected/9f27cd34-9226-4968-a4e7-83788b460159-kube-api-access-gzdhh\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.469901 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-service-ca-bundle\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470043 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470104 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470132 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470155 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470199 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470281 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-serving-cert\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470347 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470374 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a9c53b8-81fc-4e05-bb9b-f205aea24e47-metrics-tls\") pod \"dns-operator-744455d44c-h6d8x\" (UID: \"5a9c53b8-81fc-4e05-bb9b-f205aea24e47\") " pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470397 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-image-import-ca\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470510 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-config\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470613 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470642 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f27cd34-9226-4968-a4e7-83788b460159-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470762 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7vcw\" (UniqueName: \"kubernetes.io/projected/e3bd54a5-561b-4207-9c2b-24718a0b473f-kube-api-access-x7vcw\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470776 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470836 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-config\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470873 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb732274-6b17-47f6-ace6-485706ecd538-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.470911 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-console-config\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471019 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471157 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-etcd-client\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471183 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-client-ca\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471292 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7rv2\" (UniqueName: \"kubernetes.io/projected/bb732274-6b17-47f6-ace6-485706ecd538-kube-api-access-f7rv2\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471375 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471408 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-config\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.471608 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.472217 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-client-ca\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.472448 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-config\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.472485 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/74542077-889e-48b4-9af3-3f0ffde94fc8-image-import-ca\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.472821 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-serving-cert\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.472941 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.473045 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.473172 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-serving-cert\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.473179 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.473615 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.474710 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.474360 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cc929"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.474892 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.474783 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.476437 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cc929" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.477669 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.478432 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b7bb127-9196-4baa-b581-5ec01aa6da33-serving-cert\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.478538 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-etcd-client\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.478699 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/74542077-889e-48b4-9af3-3f0ffde94fc8-encryption-config\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.478858 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.479050 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.479453 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5bsgf"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.481011 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.482210 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5wfhh"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.482598 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-serving-cert\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.483466 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.484519 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.485638 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vbkql"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.486817 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.487925 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qwjms"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.489051 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.490276 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fkw88"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.491377 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.492530 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.498808 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.507366 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.512089 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.516066 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csp56"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.516530 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.520169 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.522202 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.523859 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.525227 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qzm9p"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.526626 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.526664 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.528989 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pwx2p"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.530446 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.532746 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cc929"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.534423 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.535108 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.535437 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.536869 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qzm9p"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.538539 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.539926 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dgmfk"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.541702 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-67g6s"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.543277 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-stbh4"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.544634 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-g2jss"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.545374 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.546138 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-thsh7"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.546975 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.547883 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-g2jss"] Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.555692 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574051 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8680c4-28b9-4242-8249-ad3efcd042a0-config\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574123 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4a8680c4-28b9-4242-8249-ad3efcd042a0-auth-proxy-config\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-trusted-ca-bundle\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574215 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xddmp\" (UniqueName: \"kubernetes.io/projected/9c53cf1a-a171-429e-b852-27132e72cc90-kube-api-access-xddmp\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574249 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddq8t\" (UniqueName: \"kubernetes.io/projected/de406636-3ee1-4b50-b102-8167b8a87963-kube-api-access-ddq8t\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574405 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574480 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1cac92c-afb8-4309-9fb8-28cd2f334abd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574509 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkbns\" (UniqueName: \"kubernetes.io/projected/c2582dcb-a8bc-4c67-8e7f-43c1303112d6-kube-api-access-rkbns\") pod \"package-server-manager-789f6589d5-gsffr\" (UID: \"c2582dcb-a8bc-4c67-8e7f-43c1303112d6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574601 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9c53cf1a-a171-429e-b852-27132e72cc90-srv-cert\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574656 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cd25b52b-0498-4bb0-a818-a674ab7b0eee-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574683 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-encryption-config\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574741 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7rjp\" (UniqueName: \"kubernetes.io/projected/cc835cc1-b1bd-472d-9839-85427e5a6ad7-kube-api-access-w7rjp\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574800 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-service-ca-bundle\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574838 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3bd54a5-561b-4207-9c2b-24718a0b473f-serving-cert\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574900 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-serving-cert\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574924 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-config\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.574980 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37364f76-6a05-44ef-b24d-7fc0cd69145d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5056ac5a-1bc7-4298-9417-995c16c28ec1-secret-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575057 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575066 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-trusted-ca\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575124 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzdhh\" (UniqueName: \"kubernetes.io/projected/9f27cd34-9226-4968-a4e7-83788b460159-kube-api-access-gzdhh\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575151 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-metrics-certs\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575169 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bc70b2d-20e0-4336-acf6-8606980f9b87-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575190 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ca6bec4f-27fb-4c21-a7cc-b4611385dc27-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pwx2p\" (UID: \"ca6bec4f-27fb-4c21-a7cc-b4611385dc27\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575208 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a9c53b8-81fc-4e05-bb9b-f205aea24e47-metrics-tls\") pod \"dns-operator-744455d44c-h6d8x\" (UID: \"5a9c53b8-81fc-4e05-bb9b-f205aea24e47\") " pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575217 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-trusted-ca-bundle\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575230 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbc9f\" (UniqueName: \"kubernetes.io/projected/16cbfdaf-1b4c-463e-9767-6879dc885056-kube-api-access-fbc9f\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575389 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575402 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd25b52b-0498-4bb0-a818-a674ab7b0eee-proxy-tls\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575426 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7rv2\" (UniqueName: \"kubernetes.io/projected/bb732274-6b17-47f6-ace6-485706ecd538-kube-api-access-f7rv2\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575443 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575476 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc835cc1-b1bd-472d-9839-85427e5a6ad7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575494 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a493da47-cbcd-4a06-bd49-d92edf65902b-srv-cert\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575511 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-default-certificate\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575532 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4tp9\" (UniqueName: \"kubernetes.io/projected/ca6bec4f-27fb-4c21-a7cc-b4611385dc27-kube-api-access-k4tp9\") pod \"multus-admission-controller-857f4d67dd-pwx2p\" (UID: \"ca6bec4f-27fb-4c21-a7cc-b4611385dc27\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575556 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3bd54a5-561b-4207-9c2b-24718a0b473f-config\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575573 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cspdf\" (UniqueName: \"kubernetes.io/projected/307823df-a4bd-4d96-a33c-f1684b6a00a6-kube-api-access-cspdf\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575590 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pgdr\" (UniqueName: \"kubernetes.io/projected/37364f76-6a05-44ef-b24d-7fc0cd69145d-kube-api-access-2pgdr\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575608 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1b73cd49-492e-443e-b199-b7f9b80b066c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c54gv\" (UID: \"1b73cd49-492e-443e-b199-b7f9b80b066c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575627 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2chqc\" (UniqueName: \"kubernetes.io/projected/5592da75-33af-4003-96eb-542cf6cdc256-kube-api-access-2chqc\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575650 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6ljt\" (UniqueName: \"kubernetes.io/projected/5056ac5a-1bc7-4298-9417-995c16c28ec1-kube-api-access-n6ljt\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575683 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bc70b2d-20e0-4336-acf6-8606980f9b87-config\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575711 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575740 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-metrics-tls\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575751 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8680c4-28b9-4242-8249-ad3efcd042a0-config\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575769 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbsr5\" (UniqueName: \"kubernetes.io/projected/82926c75-1f3c-4230-b981-f341f4e17ee2-kube-api-access-bbsr5\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575793 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cac92c-afb8-4309-9fb8-28cd2f334abd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575825 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4a8680c4-28b9-4242-8249-ad3efcd042a0-machine-approver-tls\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575882 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-ca\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575909 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2582dcb-a8bc-4c67-8e7f-43c1303112d6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gsffr\" (UID: \"c2582dcb-a8bc-4c67-8e7f-43c1303112d6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575939 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qgp9\" (UniqueName: \"kubernetes.io/projected/9c628ec2-7934-4ec0-b08c-24936a63d8fb-kube-api-access-6qgp9\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575968 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a493da47-cbcd-4a06-bd49-d92edf65902b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.575986 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f367570-d5b9-48b3-a76b-9931605c22cd-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576007 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/307823df-a4bd-4d96-a33c-f1684b6a00a6-tmpfs\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576038 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-oauth-config\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576060 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rdtn\" (UniqueName: \"kubernetes.io/projected/aa07299a-4ae2-4a48-972e-0b2be5c3db10-kube-api-access-2rdtn\") pod \"migrator-59844c95c7-thnm7\" (UID: \"aa07299a-4ae2-4a48-972e-0b2be5c3db10\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576081 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9f27cd34-9226-4968-a4e7-83788b460159-images\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576122 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cct57\" (UniqueName: \"kubernetes.io/projected/1b73cd49-492e-443e-b199-b7f9b80b066c-kube-api-access-cct57\") pod \"cluster-samples-operator-665b6dd947-c54gv\" (UID: \"1b73cd49-492e-443e-b199-b7f9b80b066c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576140 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576164 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576191 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/307823df-a4bd-4d96-a33c-f1684b6a00a6-webhook-cert\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576216 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-service-ca\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576235 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxrkf\" (UniqueName: \"kubernetes.io/projected/cd25b52b-0498-4bb0-a818-a674ab7b0eee-kube-api-access-dxrkf\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576259 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/749499f8-a6c6-403c-9a4a-26242d44b3c5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576322 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4ct8\" (UniqueName: \"kubernetes.io/projected/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-kube-api-access-q4ct8\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576349 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576367 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-stats-auth\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576386 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f27cd34-9226-4968-a4e7-83788b460159-config\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576402 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/749499f8-a6c6-403c-9a4a-26242d44b3c5-config\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576426 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/749499f8-a6c6-403c-9a4a-26242d44b3c5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576442 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvqkn\" (UniqueName: \"kubernetes.io/projected/4a8680c4-28b9-4242-8249-ad3efcd042a0-kube-api-access-lvqkn\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576461 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c628ec2-7934-4ec0-b08c-24936a63d8fb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576478 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3bd54a5-561b-4207-9c2b-24718a0b473f-trusted-ca\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576495 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37364f76-6a05-44ef-b24d-7fc0cd69145d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576518 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pbx\" (UniqueName: \"kubernetes.io/projected/a27662f3-0c05-4afb-addc-ebcfdf5a6b95-kube-api-access-p4pbx\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkjdk\" (UID: \"a27662f3-0c05-4afb-addc-ebcfdf5a6b95\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576537 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de406636-3ee1-4b50-b102-8167b8a87963-signing-cabundle\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n7fg\" (UniqueName: \"kubernetes.io/projected/a493da47-cbcd-4a06-bd49-d92edf65902b-kube-api-access-8n7fg\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576585 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc835cc1-b1bd-472d-9839-85427e5a6ad7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576606 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-serving-cert\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576621 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbmrs\" (UniqueName: \"kubernetes.io/projected/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-kube-api-access-vbmrs\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576639 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7vcw\" (UniqueName: \"kubernetes.io/projected/e3bd54a5-561b-4207-9c2b-24718a0b473f-kube-api-access-x7vcw\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576655 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f367570-d5b9-48b3-a76b-9931605c22cd-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576676 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f27cd34-9226-4968-a4e7-83788b460159-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576698 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-console-config\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576717 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb732274-6b17-47f6-ace6-485706ecd538-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576733 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c628ec2-7934-4ec0-b08c-24936a63d8fb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576749 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd25b52b-0498-4bb0-a818-a674ab7b0eee-images\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576768 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsm9s\" (UniqueName: \"kubernetes.io/projected/3f367570-d5b9-48b3-a76b-9931605c22cd-kube-api-access-rsm9s\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576769 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e3bd54a5-561b-4207-9c2b-24718a0b473f-config\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576792 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4gvd\" (UniqueName: \"kubernetes.io/projected/ee371f5c-b62b-4694-808b-f1f26a2353c5-kube-api-access-m4gvd\") pod \"downloads-7954f5f757-qwjms\" (UID: \"ee371f5c-b62b-4694-808b-f1f26a2353c5\") " pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576812 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576830 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-proxy-tls\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576849 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ncht\" (UniqueName: \"kubernetes.io/projected/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-kube-api-access-5ncht\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576867 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-service-ca\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8br8k\" (UniqueName: \"kubernetes.io/projected/5a9c53b8-81fc-4e05-bb9b-f205aea24e47-kube-api-access-8br8k\") pod \"dns-operator-744455d44c-h6d8x\" (UID: \"5a9c53b8-81fc-4e05-bb9b-f205aea24e47\") " pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576950 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb732274-6b17-47f6-ace6-485706ecd538-serving-cert\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.576975 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-serving-cert\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577045 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-etcd-client\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577072 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5592da75-33af-4003-96eb-542cf6cdc256-audit-dir\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577130 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-client\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577161 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-oauth-serving-cert\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577212 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c628ec2-7934-4ec0-b08c-24936a63d8fb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577237 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a27662f3-0c05-4afb-addc-ebcfdf5a6b95-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkjdk\" (UID: \"a27662f3-0c05-4afb-addc-ebcfdf5a6b95\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577286 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/307823df-a4bd-4d96-a33c-f1684b6a00a6-apiservice-cert\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577334 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bc70b2d-20e0-4336-acf6-8606980f9b87-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577363 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c2fg\" (UniqueName: \"kubernetes.io/projected/45b72490-12d9-450f-9a00-2d467c68a35b-kube-api-access-5c2fg\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577410 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgblx\" (UniqueName: \"kubernetes.io/projected/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-kube-api-access-hgblx\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577435 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-config\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577507 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-audit-policies\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577536 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c53cf1a-a171-429e-b852-27132e72cc90-profile-collector-cert\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577589 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de406636-3ee1-4b50-b102-8167b8a87963-signing-key\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577610 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-trusted-ca\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577623 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45b72490-12d9-450f-9a00-2d467c68a35b-serving-cert\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577698 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cac92c-afb8-4309-9fb8-28cd2f334abd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.577987 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-console-config\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.578427 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-config\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.578685 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-encryption-config\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.578833 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9f27cd34-9226-4968-a4e7-83788b460159-config\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.579629 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4a8680c4-28b9-4242-8249-ad3efcd042a0-auth-proxy-config\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.579776 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3bd54a5-561b-4207-9c2b-24718a0b473f-trusted-ca\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.579776 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.579868 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/749499f8-a6c6-403c-9a4a-26242d44b3c5-config\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.580010 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-ca\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.580135 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5592da75-33af-4003-96eb-542cf6cdc256-audit-dir\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.580220 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e3bd54a5-561b-4207-9c2b-24718a0b473f-serving-cert\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.581085 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5592da75-33af-4003-96eb-542cf6cdc256-audit-policies\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.581392 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9f27cd34-9226-4968-a4e7-83788b460159-images\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.581492 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4a8680c4-28b9-4242-8249-ad3efcd042a0-machine-approver-tls\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.581748 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9f27cd34-9226-4968-a4e7-83788b460159-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.582104 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-service-ca\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.582477 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-oauth-serving-cert\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.582679 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb732274-6b17-47f6-ace6-485706ecd538-serving-cert\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.582735 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bb732274-6b17-47f6-ace6-485706ecd538-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.582743 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-service-ca\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.582975 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45b72490-12d9-450f-9a00-2d467c68a35b-serving-cert\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.584025 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45b72490-12d9-450f-9a00-2d467c68a35b-etcd-client\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.584468 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-metrics-tls\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.584477 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-etcd-client\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.584989 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-oauth-config\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.585200 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/749499f8-a6c6-403c-9a4a-26242d44b3c5-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.585868 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/5a9c53b8-81fc-4e05-bb9b-f205aea24e47-metrics-tls\") pod \"dns-operator-744455d44c-h6d8x\" (UID: \"5a9c53b8-81fc-4e05-bb9b-f205aea24e47\") " pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.587468 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/1b73cd49-492e-443e-b199-b7f9b80b066c-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c54gv\" (UID: \"1b73cd49-492e-443e-b199-b7f9b80b066c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.588036 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-serving-cert\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.595760 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.599940 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5592da75-33af-4003-96eb-542cf6cdc256-serving-cert\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.615966 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.635090 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.656091 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.675066 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679763 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2582dcb-a8bc-4c67-8e7f-43c1303112d6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gsffr\" (UID: \"c2582dcb-a8bc-4c67-8e7f-43c1303112d6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679816 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qgp9\" (UniqueName: \"kubernetes.io/projected/9c628ec2-7934-4ec0-b08c-24936a63d8fb-kube-api-access-6qgp9\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679845 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a493da47-cbcd-4a06-bd49-d92edf65902b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679863 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/307823df-a4bd-4d96-a33c-f1684b6a00a6-tmpfs\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679880 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f367570-d5b9-48b3-a76b-9931605c22cd-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679926 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rdtn\" (UniqueName: \"kubernetes.io/projected/aa07299a-4ae2-4a48-972e-0b2be5c3db10-kube-api-access-2rdtn\") pod \"migrator-59844c95c7-thnm7\" (UID: \"aa07299a-4ae2-4a48-972e-0b2be5c3db10\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679958 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.679981 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/307823df-a4bd-4d96-a33c-f1684b6a00a6-webhook-cert\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680000 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxrkf\" (UniqueName: \"kubernetes.io/projected/cd25b52b-0498-4bb0-a818-a674ab7b0eee-kube-api-access-dxrkf\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680020 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680042 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4ct8\" (UniqueName: \"kubernetes.io/projected/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-kube-api-access-q4ct8\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680067 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-stats-auth\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680103 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c628ec2-7934-4ec0-b08c-24936a63d8fb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680119 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37364f76-6a05-44ef-b24d-7fc0cd69145d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680139 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4pbx\" (UniqueName: \"kubernetes.io/projected/a27662f3-0c05-4afb-addc-ebcfdf5a6b95-kube-api-access-p4pbx\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkjdk\" (UID: \"a27662f3-0c05-4afb-addc-ebcfdf5a6b95\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de406636-3ee1-4b50-b102-8167b8a87963-signing-cabundle\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680182 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n7fg\" (UniqueName: \"kubernetes.io/projected/a493da47-cbcd-4a06-bd49-d92edf65902b-kube-api-access-8n7fg\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680201 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbmrs\" (UniqueName: \"kubernetes.io/projected/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-kube-api-access-vbmrs\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680254 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc835cc1-b1bd-472d-9839-85427e5a6ad7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680271 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-serving-cert\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680307 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f367570-d5b9-48b3-a76b-9931605c22cd-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680328 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c628ec2-7934-4ec0-b08c-24936a63d8fb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680345 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd25b52b-0498-4bb0-a818-a674ab7b0eee-images\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680366 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680385 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsm9s\" (UniqueName: \"kubernetes.io/projected/3f367570-d5b9-48b3-a76b-9931605c22cd-kube-api-access-rsm9s\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680413 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-proxy-tls\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680433 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ncht\" (UniqueName: \"kubernetes.io/projected/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-kube-api-access-5ncht\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680480 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/307823df-a4bd-4d96-a33c-f1684b6a00a6-apiservice-cert\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680501 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c628ec2-7934-4ec0-b08c-24936a63d8fb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680521 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a27662f3-0c05-4afb-addc-ebcfdf5a6b95-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkjdk\" (UID: \"a27662f3-0c05-4afb-addc-ebcfdf5a6b95\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680541 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-config\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680558 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bc70b2d-20e0-4336-acf6-8606980f9b87-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680586 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c53cf1a-a171-429e-b852-27132e72cc90-profile-collector-cert\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680606 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de406636-3ee1-4b50-b102-8167b8a87963-signing-key\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680627 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cac92c-afb8-4309-9fb8-28cd2f334abd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680645 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xddmp\" (UniqueName: \"kubernetes.io/projected/9c53cf1a-a171-429e-b852-27132e72cc90-kube-api-access-xddmp\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680662 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddq8t\" (UniqueName: \"kubernetes.io/projected/de406636-3ee1-4b50-b102-8167b8a87963-kube-api-access-ddq8t\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680683 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1cac92c-afb8-4309-9fb8-28cd2f334abd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680700 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkbns\" (UniqueName: \"kubernetes.io/projected/c2582dcb-a8bc-4c67-8e7f-43c1303112d6-kube-api-access-rkbns\") pod \"package-server-manager-789f6589d5-gsffr\" (UID: \"c2582dcb-a8bc-4c67-8e7f-43c1303112d6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680696 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/307823df-a4bd-4d96-a33c-f1684b6a00a6-tmpfs\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680716 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9c53cf1a-a171-429e-b852-27132e72cc90-srv-cert\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680822 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cd25b52b-0498-4bb0-a818-a674ab7b0eee-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680883 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7rjp\" (UniqueName: \"kubernetes.io/projected/cc835cc1-b1bd-472d-9839-85427e5a6ad7-kube-api-access-w7rjp\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680914 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-service-ca-bundle\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680959 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37364f76-6a05-44ef-b24d-7fc0cd69145d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.680986 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5056ac5a-1bc7-4298-9417-995c16c28ec1-secret-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681038 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bc70b2d-20e0-4336-acf6-8606980f9b87-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681067 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-metrics-certs\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681091 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ca6bec4f-27fb-4c21-a7cc-b4611385dc27-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pwx2p\" (UID: \"ca6bec4f-27fb-4c21-a7cc-b4611385dc27\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681124 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbc9f\" (UniqueName: \"kubernetes.io/projected/16cbfdaf-1b4c-463e-9767-6879dc885056-kube-api-access-fbc9f\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681203 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681269 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd25b52b-0498-4bb0-a818-a674ab7b0eee-proxy-tls\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681341 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-default-certificate\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681368 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc835cc1-b1bd-472d-9839-85427e5a6ad7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681390 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a493da47-cbcd-4a06-bd49-d92edf65902b-srv-cert\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681417 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pgdr\" (UniqueName: \"kubernetes.io/projected/37364f76-6a05-44ef-b24d-7fc0cd69145d-kube-api-access-2pgdr\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4tp9\" (UniqueName: \"kubernetes.io/projected/ca6bec4f-27fb-4c21-a7cc-b4611385dc27-kube-api-access-k4tp9\") pod \"multus-admission-controller-857f4d67dd-pwx2p\" (UID: \"ca6bec4f-27fb-4c21-a7cc-b4611385dc27\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681463 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cspdf\" (UniqueName: \"kubernetes.io/projected/307823df-a4bd-4d96-a33c-f1684b6a00a6-kube-api-access-cspdf\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681495 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6ljt\" (UniqueName: \"kubernetes.io/projected/5056ac5a-1bc7-4298-9417-995c16c28ec1-kube-api-access-n6ljt\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681534 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bc70b2d-20e0-4336-acf6-8606980f9b87-config\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681577 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cac92c-afb8-4309-9fb8-28cd2f334abd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681680 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1cac92c-afb8-4309-9fb8-28cd2f334abd-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.681712 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cd25b52b-0498-4bb0-a818-a674ab7b0eee-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.682411 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.683710 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e1cac92c-afb8-4309-9fb8-28cd2f334abd-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.684403 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/a27662f3-0c05-4afb-addc-ebcfdf5a6b95-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkjdk\" (UID: \"a27662f3-0c05-4afb-addc-ebcfdf5a6b95\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.695881 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.711829 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/9c53cf1a-a171-429e-b852-27132e72cc90-srv-cert\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.716568 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.723982 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/a493da47-cbcd-4a06-bd49-d92edf65902b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.724171 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/9c53cf1a-a171-429e-b852-27132e72cc90-profile-collector-cert\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.725049 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5056ac5a-1bc7-4298-9417-995c16c28ec1-secret-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.735026 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.755133 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.776455 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.796121 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.816205 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.824639 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f367570-d5b9-48b3-a76b-9931605c22cd-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.843076 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.852247 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9c628ec2-7934-4ec0-b08c-24936a63d8fb-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.855214 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.861510 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f367570-d5b9-48b3-a76b-9931605c22cd-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.875841 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.895636 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.915510 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.924022 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c628ec2-7934-4ec0-b08c-24936a63d8fb-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.935166 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.954919 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.974778 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.985262 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ca6bec4f-27fb-4c21-a7cc-b4611385dc27-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-pwx2p\" (UID: \"ca6bec4f-27fb-4c21-a7cc-b4611385dc27\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:09 crc kubenswrapper[4979]: I1204 11:45:09.995955 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.006355 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cc835cc1-b1bd-472d-9839-85427e5a6ad7-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.015712 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.035574 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.042362 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/cd25b52b-0498-4bb0-a818-a674ab7b0eee-images\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.056072 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.061746 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc835cc1-b1bd-472d-9839-85427e5a6ad7-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.075905 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.096263 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.116330 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.124734 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cd25b52b-0498-4bb0-a818-a674ab7b0eee-proxy-tls\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.154972 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.165659 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-default-certificate\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.175242 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.184392 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-stats-auth\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.195350 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.217404 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.235466 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.243086 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-service-ca-bundle\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.255576 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.264882 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-metrics-certs\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.275466 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.295881 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.315410 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.335353 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.355657 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.376081 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.385835 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-proxy-tls\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.395782 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.406038 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/a493da47-cbcd-4a06-bd49-d92edf65902b-srv-cert\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.416056 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.425289 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/307823df-a4bd-4d96-a33c-f1684b6a00a6-webhook-cert\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.425287 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/307823df-a4bd-4d96-a33c-f1684b6a00a6-apiservice-cert\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.444717 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.453218 4979 request.go:700] Waited for 1.016952577s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-dockercfg-5nsgg&limit=500&resourceVersion=0 Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.455459 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.457576 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.485180 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.494111 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.495603 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.515845 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.536101 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.543875 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c2582dcb-a8bc-4c67-8e7f-43c1303112d6-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gsffr\" (UID: \"c2582dcb-a8bc-4c67-8e7f-43c1303112d6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.554960 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.575707 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.595762 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.614919 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.636384 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.655352 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.664193 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bc70b2d-20e0-4336-acf6-8606980f9b87-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.676074 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680461 4979 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680513 4979 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680562 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-serving-cert podName:a2ba54a7-a273-41ce-b7a1-7dc4949a94d5 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.180542154 +0000 UTC m=+135.454837958 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-serving-cert") pod "service-ca-operator-777779d784-67g6s" (UID: "a2ba54a7-a273-41ce-b7a1-7dc4949a94d5") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680563 4979 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680576 4979 secret.go:188] Couldn't get secret openshift-apiserver-operator/openshift-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680582 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/de406636-3ee1-4b50-b102-8167b8a87963-signing-cabundle podName:de406636-3ee1-4b50-b102-8167b8a87963 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.180574625 +0000 UTC m=+135.454870429 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/de406636-3ee1-4b50-b102-8167b8a87963-signing-cabundle") pod "service-ca-9c57cc56f-dgmfk" (UID: "de406636-3ee1-4b50-b102-8167b8a87963") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680637 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume podName:5056ac5a-1bc7-4298-9417-995c16c28ec1 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.180612876 +0000 UTC m=+135.454908740 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume") pod "collect-profiles-29414145-4glpc" (UID: "5056ac5a-1bc7-4298-9417-995c16c28ec1") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680644 4979 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680656 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37364f76-6a05-44ef-b24d-7fc0cd69145d-serving-cert podName:37364f76-6a05-44ef-b24d-7fc0cd69145d nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.180647027 +0000 UTC m=+135.454942951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/37364f76-6a05-44ef-b24d-7fc0cd69145d-serving-cert") pod "openshift-apiserver-operator-796bbdcf4f-99f98" (UID: "37364f76-6a05-44ef-b24d-7fc0cd69145d") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680676 4979 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680683 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-config podName:a2ba54a7-a273-41ce-b7a1-7dc4949a94d5 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.180668647 +0000 UTC m=+135.454964521 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-config") pod "service-ca-operator-777779d784-67g6s" (UID: "a2ba54a7-a273-41ce-b7a1-7dc4949a94d5") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.680707 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/de406636-3ee1-4b50-b102-8167b8a87963-signing-key podName:de406636-3ee1-4b50-b102-8167b8a87963 nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.180698228 +0000 UTC m=+135.454994142 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/de406636-3ee1-4b50-b102-8167b8a87963-signing-key") pod "service-ca-9c57cc56f-dgmfk" (UID: "de406636-3ee1-4b50-b102-8167b8a87963") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.681117 4979 configmap.go:193] Couldn't get configMap openshift-apiserver-operator/openshift-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: E1204 11:45:10.681257 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/37364f76-6a05-44ef-b24d-7fc0cd69145d-config podName:37364f76-6a05-44ef-b24d-7fc0cd69145d nodeName:}" failed. No retries permitted until 2025-12-04 11:45:11.181243483 +0000 UTC m=+135.455539287 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/37364f76-6a05-44ef-b24d-7fc0cd69145d-config") pod "openshift-apiserver-operator-796bbdcf4f-99f98" (UID: "37364f76-6a05-44ef-b24d-7fc0cd69145d") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.682486 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bc70b2d-20e0-4336-acf6-8606980f9b87-config\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.694856 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.715575 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.735870 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.755913 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.775961 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.795491 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.816024 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.836205 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.855340 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.874780 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.896740 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.916074 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.935843 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.956435 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.975679 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 11:45:10 crc kubenswrapper[4979]: I1204 11:45:10.995733 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.015135 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.050095 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv4n9\" (UniqueName: \"kubernetes.io/projected/6b7bb127-9196-4baa-b581-5ec01aa6da33-kube-api-access-tv4n9\") pod \"controller-manager-879f6c89f-bhnlh\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.068750 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr6kb\" (UniqueName: \"kubernetes.io/projected/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-kube-api-access-kr6kb\") pod \"route-controller-manager-6576b87f9c-mdvl2\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.093720 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-424kk\" (UniqueName: \"kubernetes.io/projected/e31c1aff-c394-476a-888e-c0e877f59717-kube-api-access-424kk\") pod \"oauth-openshift-558db77b4-4v569\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.102819 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.122431 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vkpc\" (UniqueName: \"kubernetes.io/projected/d4ab823d-d6ba-4e56-8c79-932a23b7aa36-kube-api-access-6vkpc\") pod \"authentication-operator-69f744f599-fkw88\" (UID: \"d4ab823d-d6ba-4e56-8c79-932a23b7aa36\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.139863 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb6l5\" (UniqueName: \"kubernetes.io/projected/74542077-889e-48b4-9af3-3f0ffde94fc8-kube-api-access-xb6l5\") pod \"apiserver-76f77b778f-dkvxc\" (UID: \"74542077-889e-48b4-9af3-3f0ffde94fc8\") " pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.155675 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.164186 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.175585 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.188154 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.196324 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.201715 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209140 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209224 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de406636-3ee1-4b50-b102-8167b8a87963-signing-cabundle\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209261 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37364f76-6a05-44ef-b24d-7fc0cd69145d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209335 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-serving-cert\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209450 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-config\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209476 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de406636-3ee1-4b50-b102-8167b8a87963-signing-key\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.209535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37364f76-6a05-44ef-b24d-7fc0cd69145d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.210902 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37364f76-6a05-44ef-b24d-7fc0cd69145d-config\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.211764 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-config\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.212494 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.213226 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de406636-3ee1-4b50-b102-8167b8a87963-signing-cabundle\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.215219 4979 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.218542 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37364f76-6a05-44ef-b24d-7fc0cd69145d-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.220443 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de406636-3ee1-4b50-b102-8167b8a87963-signing-key\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.220673 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-serving-cert\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.235583 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.254872 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.275675 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.295238 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.315774 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.335256 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhnlh"] Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.336013 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 11:45:11 crc kubenswrapper[4979]: W1204 11:45:11.346670 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b7bb127_9196_4baa_b581_5ec01aa6da33.slice/crio-c71b7b3f12fe147f15593e4c52da47276281e51c392589c45517edc06fc9e94e WatchSource:0}: Error finding container c71b7b3f12fe147f15593e4c52da47276281e51c392589c45517edc06fc9e94e: Status 404 returned error can't find the container with id c71b7b3f12fe147f15593e4c52da47276281e51c392589c45517edc06fc9e94e Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.354806 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.375860 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.382355 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2"] Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.396909 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 11:45:11 crc kubenswrapper[4979]: W1204 11:45:11.405492 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17eb3c3d_c64c_416e_aa5f_c6839db5d1fc.slice/crio-276980da103f03402f3dd3d08dfeecb7b182abceb049c8ec29f7cb1c1782f329 WatchSource:0}: Error finding container 276980da103f03402f3dd3d08dfeecb7b182abceb049c8ec29f7cb1c1782f329: Status 404 returned error can't find the container with id 276980da103f03402f3dd3d08dfeecb7b182abceb049c8ec29f7cb1c1782f329 Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.409407 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4v569"] Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.428501 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.433784 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzdhh\" (UniqueName: \"kubernetes.io/projected/9f27cd34-9226-4968-a4e7-83788b460159-kube-api-access-gzdhh\") pod \"machine-api-operator-5694c8668f-5bsgf\" (UID: \"9f27cd34-9226-4968-a4e7-83788b460159\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.450585 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7rv2\" (UniqueName: \"kubernetes.io/projected/bb732274-6b17-47f6-ace6-485706ecd538-kube-api-access-f7rv2\") pod \"openshift-config-operator-7777fb866f-fnchd\" (UID: \"bb732274-6b17-47f6-ace6-485706ecd538\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.459217 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fkw88"] Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.471580 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2chqc\" (UniqueName: \"kubernetes.io/projected/5592da75-33af-4003-96eb-542cf6cdc256-kube-api-access-2chqc\") pod \"apiserver-7bbb656c7d-r59vv\" (UID: \"5592da75-33af-4003-96eb-542cf6cdc256\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:11 crc kubenswrapper[4979]: W1204 11:45:11.472052 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4ab823d_d6ba_4e56_8c79_932a23b7aa36.slice/crio-e03fbf2f1ccbb6cef095f0ffcce9820e1876fe0a4948a092dca0d714389053c9 WatchSource:0}: Error finding container e03fbf2f1ccbb6cef095f0ffcce9820e1876fe0a4948a092dca0d714389053c9: Status 404 returned error can't find the container with id e03fbf2f1ccbb6cef095f0ffcce9820e1876fe0a4948a092dca0d714389053c9 Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.472933 4979 request.go:700] Waited for 1.894130702s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.491566 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbsr5\" (UniqueName: \"kubernetes.io/projected/82926c75-1f3c-4230-b981-f341f4e17ee2-kube-api-access-bbsr5\") pod \"console-f9d7485db-5wfhh\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.517445 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7vcw\" (UniqueName: \"kubernetes.io/projected/e3bd54a5-561b-4207-9c2b-24718a0b473f-kube-api-access-x7vcw\") pod \"console-operator-58897d9998-xmjpc\" (UID: \"e3bd54a5-561b-4207-9c2b-24718a0b473f\") " pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.522733 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.530514 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.540991 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvqkn\" (UniqueName: \"kubernetes.io/projected/4a8680c4-28b9-4242-8249-ad3efcd042a0-kube-api-access-lvqkn\") pod \"machine-approver-56656f9798-m55jc\" (UID: \"4a8680c4-28b9-4242-8249-ad3efcd042a0\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.546136 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.552174 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/749499f8-a6c6-403c-9a4a-26242d44b3c5-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-x46bj\" (UID: \"749499f8-a6c6-403c-9a4a-26242d44b3c5\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.567988 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4gvd\" (UniqueName: \"kubernetes.io/projected/ee371f5c-b62b-4694-808b-f1f26a2353c5-kube-api-access-m4gvd\") pod \"downloads-7954f5f757-qwjms\" (UID: \"ee371f5c-b62b-4694-808b-f1f26a2353c5\") " pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.575866 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.611964 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.612370 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cct57\" (UniqueName: \"kubernetes.io/projected/1b73cd49-492e-443e-b199-b7f9b80b066c-kube-api-access-cct57\") pod \"cluster-samples-operator-665b6dd947-c54gv\" (UID: \"1b73cd49-492e-443e-b199-b7f9b80b066c\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.618364 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.629145 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dkvxc"] Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.630920 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-bound-sa-token\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.636661 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.645589 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.671338 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c2fg\" (UniqueName: \"kubernetes.io/projected/45b72490-12d9-450f-9a00-2d467c68a35b-kube-api-access-5c2fg\") pod \"etcd-operator-b45778765-vbkql\" (UID: \"45b72490-12d9-450f-9a00-2d467c68a35b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.690557 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qgp9\" (UniqueName: \"kubernetes.io/projected/9c628ec2-7934-4ec0-b08c-24936a63d8fb-kube-api-access-6qgp9\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.711252 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rdtn\" (UniqueName: \"kubernetes.io/projected/aa07299a-4ae2-4a48-972e-0b2be5c3db10-kube-api-access-2rdtn\") pod \"migrator-59844c95c7-thnm7\" (UID: \"aa07299a-4ae2-4a48-972e-0b2be5c3db10\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.733845 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxrkf\" (UniqueName: \"kubernetes.io/projected/cd25b52b-0498-4bb0-a818-a674ab7b0eee-kube-api-access-dxrkf\") pod \"machine-config-operator-74547568cd-vdmr2\" (UID: \"cd25b52b-0498-4bb0-a818-a674ab7b0eee\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.749445 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.753793 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4ct8\" (UniqueName: \"kubernetes.io/projected/dfb0a0f6-642a-469a-891f-5d7bb9a76b01-kube-api-access-q4ct8\") pod \"machine-config-controller-84d6567774-q5vs8\" (UID: \"dfb0a0f6-642a-469a-891f-5d7bb9a76b01\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.765437 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.770621 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n7fg\" (UniqueName: \"kubernetes.io/projected/a493da47-cbcd-4a06-bd49-d92edf65902b-kube-api-access-8n7fg\") pod \"olm-operator-6b444d44fb-tcrt4\" (UID: \"a493da47-cbcd-4a06-bd49-d92edf65902b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.771040 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.782685 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.788953 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbmrs\" (UniqueName: \"kubernetes.io/projected/a2ba54a7-a273-41ce-b7a1-7dc4949a94d5-kube-api-access-vbmrs\") pod \"service-ca-operator-777779d784-67g6s\" (UID: \"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.810546 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4pbx\" (UniqueName: \"kubernetes.io/projected/a27662f3-0c05-4afb-addc-ebcfdf5a6b95-kube-api-access-p4pbx\") pod \"control-plane-machine-set-operator-78cbb6b69f-dkjdk\" (UID: \"a27662f3-0c05-4afb-addc-ebcfdf5a6b95\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.830644 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9c628ec2-7934-4ec0-b08c-24936a63d8fb-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-64jzx\" (UID: \"9c628ec2-7934-4ec0-b08c-24936a63d8fb\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.837764 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.849590 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsm9s\" (UniqueName: \"kubernetes.io/projected/3f367570-d5b9-48b3-a76b-9931605c22cd-kube-api-access-rsm9s\") pod \"openshift-controller-manager-operator-756b6f6bc6-29vsx\" (UID: \"3f367570-d5b9-48b3-a76b-9931605c22cd\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.854136 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.868516 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ncht\" (UniqueName: \"kubernetes.io/projected/abc2fd1c-2098-4d1e-b86e-9c4e8be715ca-kube-api-access-5ncht\") pod \"router-default-5444994796-j84bh\" (UID: \"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca\") " pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.872093 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.893611 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xddmp\" (UniqueName: \"kubernetes.io/projected/9c53cf1a-a171-429e-b852-27132e72cc90-kube-api-access-xddmp\") pod \"catalog-operator-68c6474976-c6fzf\" (UID: \"9c53cf1a-a171-429e-b852-27132e72cc90\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.925902 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkbns\" (UniqueName: \"kubernetes.io/projected/c2582dcb-a8bc-4c67-8e7f-43c1303112d6-kube-api-access-rkbns\") pod \"package-server-manager-789f6589d5-gsffr\" (UID: \"c2582dcb-a8bc-4c67-8e7f-43c1303112d6\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.941408 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddq8t\" (UniqueName: \"kubernetes.io/projected/de406636-3ee1-4b50-b102-8167b8a87963-kube-api-access-ddq8t\") pod \"service-ca-9c57cc56f-dgmfk\" (UID: \"de406636-3ee1-4b50-b102-8167b8a87963\") " pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.947784 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7rjp\" (UniqueName: \"kubernetes.io/projected/cc835cc1-b1bd-472d-9839-85427e5a6ad7-kube-api-access-w7rjp\") pod \"kube-storage-version-migrator-operator-b67b599dd-tbthq\" (UID: \"cc835cc1-b1bd-472d-9839-85427e5a6ad7\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.953681 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.970967 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3bc70b2d-20e0-4336-acf6-8606980f9b87-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-nzww8\" (UID: \"3bc70b2d-20e0-4336-acf6-8606980f9b87\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:11 crc kubenswrapper[4979]: I1204 11:45:11.972809 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:11.998630 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgblx\" (UniqueName: \"kubernetes.io/projected/3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c-kube-api-access-hgblx\") pod \"ingress-operator-5b745b69d9-cv8nc\" (UID: \"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.001812 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8br8k\" (UniqueName: \"kubernetes.io/projected/5a9c53b8-81fc-4e05-bb9b-f205aea24e47-kube-api-access-8br8k\") pod \"dns-operator-744455d44c-h6d8x\" (UID: \"5a9c53b8-81fc-4e05-bb9b-f205aea24e47\") " pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.010537 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbc9f\" (UniqueName: \"kubernetes.io/projected/16cbfdaf-1b4c-463e-9767-6879dc885056-kube-api-access-fbc9f\") pod \"marketplace-operator-79b997595-csp56\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.020689 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.024903 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4tp9\" (UniqueName: \"kubernetes.io/projected/ca6bec4f-27fb-4c21-a7cc-b4611385dc27-kube-api-access-k4tp9\") pod \"multus-admission-controller-857f4d67dd-pwx2p\" (UID: \"ca6bec4f-27fb-4c21-a7cc-b4611385dc27\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.025064 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.032425 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.034264 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6ljt\" (UniqueName: \"kubernetes.io/projected/5056ac5a-1bc7-4298-9417-995c16c28ec1-kube-api-access-n6ljt\") pod \"collect-profiles-29414145-4glpc\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.040649 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.043176 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" event={"ID":"d4ab823d-d6ba-4e56-8c79-932a23b7aa36","Type":"ContainerStarted","Data":"e03fbf2f1ccbb6cef095f0ffcce9820e1876fe0a4948a092dca0d714389053c9"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.047056 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" event={"ID":"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc","Type":"ContainerStarted","Data":"957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.047122 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" event={"ID":"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc","Type":"ContainerStarted","Data":"276980da103f03402f3dd3d08dfeecb7b182abceb049c8ec29f7cb1c1782f329"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.049519 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" event={"ID":"6b7bb127-9196-4baa-b581-5ec01aa6da33","Type":"ContainerStarted","Data":"126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.049574 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" event={"ID":"6b7bb127-9196-4baa-b581-5ec01aa6da33","Type":"ContainerStarted","Data":"c71b7b3f12fe147f15593e4c52da47276281e51c392589c45517edc06fc9e94e"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.050598 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cspdf\" (UniqueName: \"kubernetes.io/projected/307823df-a4bd-4d96-a33c-f1684b6a00a6-kube-api-access-cspdf\") pod \"packageserver-d55dfcdfc-jbsjs\" (UID: \"307823df-a4bd-4d96-a33c-f1684b6a00a6\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.053145 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" event={"ID":"74542077-889e-48b4-9af3-3f0ffde94fc8","Type":"ContainerStarted","Data":"e845776c34df309e83ffa573bbb73c74e0045547fc4964b7bd57420f4ff56a5b"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.056887 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" event={"ID":"e31c1aff-c394-476a-888e-c0e877f59717","Type":"ContainerStarted","Data":"84dcf7239b8b0cf85127fe433b4b7779d7e7226140a1ee3e82e0fbc21806918f"} Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.058473 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.070836 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cac92c-afb8-4309-9fb8-28cd2f334abd-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-9fmlb\" (UID: \"e1cac92c-afb8-4309-9fb8-28cd2f334abd\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.089510 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pgdr\" (UniqueName: \"kubernetes.io/projected/37364f76-6a05-44ef-b24d-7fc0cd69145d-kube-api-access-2pgdr\") pod \"openshift-apiserver-operator-796bbdcf4f-99f98\" (UID: \"37364f76-6a05-44ef-b24d-7fc0cd69145d\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.104402 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.111682 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.125018 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126080 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac3bc99d-9384-4e8b-b266-92aa7a417150-installation-pull-secrets\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126141 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94s4r\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-kube-api-access-94s4r\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126186 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-bound-sa-token\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126229 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-trusted-ca\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126274 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-tls\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126409 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-certificates\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126439 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac3bc99d-9384-4e8b-b266-92aa7a417150-ca-trust-extracted\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.126472 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.126892 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:12.626875842 +0000 UTC m=+136.901171646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.144178 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.152748 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.162637 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.162941 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.183840 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.225969 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.226853 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.227224 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:12.72716485 +0000 UTC m=+137.001460664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.227449 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs92p\" (UniqueName: \"kubernetes.io/projected/bf77b7e3-6944-4940-b090-bc5a1d3c2b43-kube-api-access-fs92p\") pod \"ingress-canary-g2jss\" (UID: \"bf77b7e3-6944-4940-b090-bc5a1d3c2b43\") " pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.227558 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-tls\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.227895 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-certificates\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.227976 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac3bc99d-9384-4e8b-b266-92aa7a417150-ca-trust-extracted\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228020 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228056 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf77b7e3-6944-4940-b090-bc5a1d3c2b43-cert\") pod \"ingress-canary-g2jss\" (UID: \"bf77b7e3-6944-4940-b090-bc5a1d3c2b43\") " pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228202 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac3bc99d-9384-4e8b-b266-92aa7a417150-installation-pull-secrets\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228244 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tpb7\" (UniqueName: \"kubernetes.io/projected/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-kube-api-access-9tpb7\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94s4r\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-kube-api-access-94s4r\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228652 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9cq7\" (UniqueName: \"kubernetes.io/projected/8930c76d-0bde-451d-aea5-50323292fb6b-kube-api-access-p9cq7\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228679 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bw9b\" (UniqueName: \"kubernetes.io/projected/eed6be5b-9767-4c57-9359-a9af0046c058-kube-api-access-2bw9b\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228865 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-registration-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228897 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-csi-data-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.228982 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-plugins-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.231564 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-certificates\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.233654 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac3bc99d-9384-4e8b-b266-92aa7a417150-ca-trust-extracted\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.246686 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-tls\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.246723 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-bound-sa-token\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.247001 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac3bc99d-9384-4e8b-b266-92aa7a417150-installation-pull-secrets\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.247122 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:12.747100822 +0000 UTC m=+137.021396676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.247395 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/eed6be5b-9767-4c57-9359-a9af0046c058-node-bootstrap-token\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.247419 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/eed6be5b-9767-4c57-9359-a9af0046c058-certs\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.247436 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-metrics-tls\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.247860 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-mountpoint-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.248732 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-trusted-ca\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.249283 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-socket-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.249356 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-config-volume\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.253220 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-trusted-ca\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.265056 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.277939 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-bound-sa-token\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.314865 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94s4r\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-kube-api-access-94s4r\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353594 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353757 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9cq7\" (UniqueName: \"kubernetes.io/projected/8930c76d-0bde-451d-aea5-50323292fb6b-kube-api-access-p9cq7\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353779 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bw9b\" (UniqueName: \"kubernetes.io/projected/eed6be5b-9767-4c57-9359-a9af0046c058-kube-api-access-2bw9b\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353804 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-registration-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353820 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-csi-data-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353838 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-plugins-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353870 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/eed6be5b-9767-4c57-9359-a9af0046c058-node-bootstrap-token\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353887 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/eed6be5b-9767-4c57-9359-a9af0046c058-certs\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353901 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-metrics-tls\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353917 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-mountpoint-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353937 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-socket-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.353952 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-config-volume\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.354199 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs92p\" (UniqueName: \"kubernetes.io/projected/bf77b7e3-6944-4940-b090-bc5a1d3c2b43-kube-api-access-fs92p\") pod \"ingress-canary-g2jss\" (UID: \"bf77b7e3-6944-4940-b090-bc5a1d3c2b43\") " pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.354262 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf77b7e3-6944-4940-b090-bc5a1d3c2b43-cert\") pod \"ingress-canary-g2jss\" (UID: \"bf77b7e3-6944-4940-b090-bc5a1d3c2b43\") " pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.354290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tpb7\" (UniqueName: \"kubernetes.io/projected/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-kube-api-access-9tpb7\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.354601 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:12.854580209 +0000 UTC m=+137.128876013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.355116 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-registration-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.356760 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-config-volume\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.356874 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-socket-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.355173 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-mountpoint-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.357778 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-plugins-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.357885 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/8930c76d-0bde-451d-aea5-50323292fb6b-csi-data-dir\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.368313 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/eed6be5b-9767-4c57-9359-a9af0046c058-certs\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.371617 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4"] Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.373178 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-metrics-tls\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.377709 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/eed6be5b-9767-4c57-9359-a9af0046c058-node-bootstrap-token\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.380563 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bf77b7e3-6944-4940-b090-bc5a1d3c2b43-cert\") pod \"ingress-canary-g2jss\" (UID: \"bf77b7e3-6944-4940-b090-bc5a1d3c2b43\") " pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.396676 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tpb7\" (UniqueName: \"kubernetes.io/projected/0d1cf870-bdba-419c-9534-2f8cdf10e1c7-kube-api-access-9tpb7\") pod \"dns-default-cc929\" (UID: \"0d1cf870-bdba-419c-9534-2f8cdf10e1c7\") " pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.438684 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9cq7\" (UniqueName: \"kubernetes.io/projected/8930c76d-0bde-451d-aea5-50323292fb6b-kube-api-access-p9cq7\") pod \"csi-hostpathplugin-qzm9p\" (UID: \"8930c76d-0bde-451d-aea5-50323292fb6b\") " pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.445317 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bw9b\" (UniqueName: \"kubernetes.io/projected/eed6be5b-9767-4c57-9359-a9af0046c058-kube-api-access-2bw9b\") pod \"machine-config-server-thsh7\" (UID: \"eed6be5b-9767-4c57-9359-a9af0046c058\") " pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.459910 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.460324 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:12.960284388 +0000 UTC m=+137.234580192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.477178 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv"] Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.487941 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs92p\" (UniqueName: \"kubernetes.io/projected/bf77b7e3-6944-4940-b090-bc5a1d3c2b43-kube-api-access-fs92p\") pod \"ingress-canary-g2jss\" (UID: \"bf77b7e3-6944-4940-b090-bc5a1d3c2b43\") " pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.489511 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cc929" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.530442 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.538357 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-g2jss" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.549004 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-thsh7" Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.560461 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.560685 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.060659128 +0000 UTC m=+137.334954932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.560861 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.561255 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.061244633 +0000 UTC m=+137.335540437 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: W1204 11:45:12.605502 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeed6be5b_9767_4c57_9359_a9af0046c058.slice/crio-af730c26f458105b8e6fd24030f9b3dfdf101372276cf2e1eb4205051439d0aa WatchSource:0}: Error finding container af730c26f458105b8e6fd24030f9b3dfdf101372276cf2e1eb4205051439d0aa: Status 404 returned error can't find the container with id af730c26f458105b8e6fd24030f9b3dfdf101372276cf2e1eb4205051439d0aa Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.662385 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.663170 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.163120693 +0000 UTC m=+137.437416497 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.764395 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.764867 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.264853439 +0000 UTC m=+137.539149243 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.866115 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.866911 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.366884583 +0000 UTC m=+137.641180387 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.947966 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-5wfhh"] Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.958853 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj"] Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.964349 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fnchd"] Dec 04 11:45:12 crc kubenswrapper[4979]: I1204 11:45:12.968107 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:12 crc kubenswrapper[4979]: E1204 11:45:12.968486 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.468471914 +0000 UTC m=+137.742767708 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.073251 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.073435 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.573405284 +0000 UTC m=+137.847701098 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.073669 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.074034 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.57402372 +0000 UTC m=+137.848319524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.075418 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" event={"ID":"e31c1aff-c394-476a-888e-c0e877f59717","Type":"ContainerStarted","Data":"03c276d786e6a1d3eb0c401bf51a63a1cfc2f59c29318e8e705fff1be0737f64"} Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.076745 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.077850 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-j84bh" event={"ID":"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca","Type":"ContainerStarted","Data":"57de63cfc5c5dc7baf33523218d3c257731305c78535a41cc302e59ec30f7683"} Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.079581 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-thsh7" event={"ID":"eed6be5b-9767-4c57-9359-a9af0046c058","Type":"ContainerStarted","Data":"af730c26f458105b8e6fd24030f9b3dfdf101372276cf2e1eb4205051439d0aa"} Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.080643 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" event={"ID":"a493da47-cbcd-4a06-bd49-d92edf65902b","Type":"ContainerStarted","Data":"ebe831980c0c2f4211a55638f79434137bb2e2717af5d2101aa038c1b46afb98"} Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.082115 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" event={"ID":"4a8680c4-28b9-4242-8249-ad3efcd042a0","Type":"ContainerStarted","Data":"5c52f57c9f6833fd18a0f3e5bdcfbe99ea183163dfc56522664367f9717446d1"} Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.085391 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" event={"ID":"d4ab823d-d6ba-4e56-8c79-932a23b7aa36","Type":"ContainerStarted","Data":"25cf2f99a67234c46b68b6ed5a5c5fe6d74db520008860c90537655606ce9700"} Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.085474 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.137602 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:45:13 crc kubenswrapper[4979]: W1204 11:45:13.174620 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb732274_6b17_47f6_ace6_485706ecd538.slice/crio-dd18d7f770c306d896b4b5c6c96a9143c984cdae199a77644bcd66268d717066 WatchSource:0}: Error finding container dd18d7f770c306d896b4b5c6c96a9143c984cdae199a77644bcd66268d717066: Status 404 returned error can't find the container with id dd18d7f770c306d896b4b5c6c96a9143c984cdae199a77644bcd66268d717066 Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.200644 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.202946 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.702919208 +0000 UTC m=+137.977215012 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.310332 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.311066 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.810758703 +0000 UTC m=+138.085054507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.412665 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.413164 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:13.913134626 +0000 UTC m=+138.187430430 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.515410 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.515890 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.015869367 +0000 UTC m=+138.290165221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.616775 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.616990 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.116955966 +0000 UTC m=+138.391251770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.617142 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.617511 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.117503271 +0000 UTC m=+138.391799075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.723663 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.724771 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.2247546 +0000 UTC m=+138.499050404 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.826323 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.826732 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.326714042 +0000 UTC m=+138.601009846 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.870031 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" podStartSLOduration=118.869983206 podStartE2EDuration="1m58.869983206s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:13.865871008 +0000 UTC m=+138.140166832" watchObservedRunningTime="2025-12-04 11:45:13.869983206 +0000 UTC m=+138.144279010" Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.904972 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fkw88" podStartSLOduration=119.904950932 podStartE2EDuration="1m59.904950932s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:13.903111755 +0000 UTC m=+138.177407559" watchObservedRunningTime="2025-12-04 11:45:13.904950932 +0000 UTC m=+138.179246736" Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.928427 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:13 crc kubenswrapper[4979]: E1204 11:45:13.928880 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.428856649 +0000 UTC m=+138.703152453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:13 crc kubenswrapper[4979]: I1204 11:45:13.948715 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" podStartSLOduration=119.948697658 podStartE2EDuration="1m59.948697658s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:13.946556193 +0000 UTC m=+138.220852017" watchObservedRunningTime="2025-12-04 11:45:13.948697658 +0000 UTC m=+138.222993462" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.004529 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" podStartSLOduration=120.004509511 podStartE2EDuration="2m0.004509511s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:14.003609178 +0000 UTC m=+138.277904982" watchObservedRunningTime="2025-12-04 11:45:14.004509511 +0000 UTC m=+138.278805315" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.029909 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.030438 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.53041313 +0000 UTC m=+138.804709134 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.052019 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.059088 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.083747 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-qwjms"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.091826 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf"] Dec 04 11:45:14 crc kubenswrapper[4979]: W1204 11:45:14.097829 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c53cf1a_a171_429e_b852_27132e72cc90.slice/crio-a3664cbd91a80381980af3dc798d7b4cb121b29b837eb82537e90dffff437854 WatchSource:0}: Error finding container a3664cbd91a80381980af3dc798d7b4cb121b29b837eb82537e90dffff437854: Status 404 returned error can't find the container with id a3664cbd91a80381980af3dc798d7b4cb121b29b837eb82537e90dffff437854 Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.123561 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" event={"ID":"cd25b52b-0498-4bb0-a818-a674ab7b0eee","Type":"ContainerStarted","Data":"8b74af6315043e3b9085d1fe59101e731e701a1a12c7cf043a222f705b52e91f"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.126556 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-thsh7" event={"ID":"eed6be5b-9767-4c57-9359-a9af0046c058","Type":"ContainerStarted","Data":"01695a5cd603edeb6f46b067591ab54a1558c4e9454cbffc31cc1bce8d1a6b05"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.132673 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.132847 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.632828494 +0000 UTC m=+138.907124298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.133088 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.133609 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.633600463 +0000 UTC m=+138.907896267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.144606 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" event={"ID":"a493da47-cbcd-4a06-bd49-d92edf65902b","Type":"ContainerStarted","Data":"f1a99c1dbdb376e6564cb9aae77ebe04396a8bd6e9c92ade9e9abef317f34692"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.145488 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.146193 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.147147 4979 generic.go:334] "Generic (PLEG): container finished" podID="74542077-889e-48b4-9af3-3f0ffde94fc8" containerID="a692b41e0fbc063b35f229d6e8b078fd9eaf76e39a9b6360c28ea6ebf3150d75" exitCode=0 Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.147219 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" event={"ID":"74542077-889e-48b4-9af3-3f0ffde94fc8","Type":"ContainerDied","Data":"a692b41e0fbc063b35f229d6e8b078fd9eaf76e39a9b6360c28ea6ebf3150d75"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.171160 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-thsh7" podStartSLOduration=5.171138917 podStartE2EDuration="5.171138917s" podCreationTimestamp="2025-12-04 11:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:14.170729167 +0000 UTC m=+138.445024971" watchObservedRunningTime="2025-12-04 11:45:14.171138917 +0000 UTC m=+138.445434731" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.188351 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.238962 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.239870 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.739849447 +0000 UTC m=+139.014145251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.249354 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5wfhh" event={"ID":"82926c75-1f3c-4230-b981-f341f4e17ee2","Type":"ContainerStarted","Data":"0f0a2eb21f11096a22e4f334377e853462c11e04474f9bf7e1a383103c687db8"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.293012 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" event={"ID":"749499f8-a6c6-403c-9a4a-26242d44b3c5","Type":"ContainerStarted","Data":"75d60463e7736de1e039250b63756fa88e8f7135cd7f38d55a011f6355147f5c"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.320641 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" event={"ID":"1b73cd49-492e-443e-b199-b7f9b80b066c","Type":"ContainerStarted","Data":"4dbf13eadc8eece86437fccfefee9a9fe529b0de3b7db3034351b7d0a49e76ec"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.326720 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-j84bh" event={"ID":"abc2fd1c-2098-4d1e-b86e-9c4e8be715ca","Type":"ContainerStarted","Data":"95bd7e50007c01060428c00a7f9a21c2e713e0191fe4b854b5766d6efcf361c4"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.340915 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.341506 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.841483181 +0000 UTC m=+139.115779055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.343135 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tcrt4" podStartSLOduration=119.343118694 podStartE2EDuration="1m59.343118694s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:14.269964417 +0000 UTC m=+138.544260241" watchObservedRunningTime="2025-12-04 11:45:14.343118694 +0000 UTC m=+138.617414498" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.356010 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-xmjpc"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.364647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" event={"ID":"4a8680c4-28b9-4242-8249-ad3efcd042a0","Type":"ContainerStarted","Data":"00385ca1dc2a3aaea6dc606d29297e954f3ac68b15156d9b58f64da30d092483"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.401367 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.401532 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.401547 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" event={"ID":"bb732274-6b17-47f6-ace6-485706ecd538","Type":"ContainerStarted","Data":"dd18d7f770c306d896b4b5c6c96a9143c984cdae199a77644bcd66268d717066"} Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.413938 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.423265 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8"] Dec 04 11:45:14 crc kubenswrapper[4979]: W1204 11:45:14.435820 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bc70b2d_20e0_4336_acf6_8606980f9b87.slice/crio-7f4a422f29401a80ccf7977022f24ac381bd89912f3ca099871951dc05d0006c WatchSource:0}: Error finding container 7f4a422f29401a80ccf7977022f24ac381bd89912f3ca099871951dc05d0006c: Status 404 returned error can't find the container with id 7f4a422f29401a80ccf7977022f24ac381bd89912f3ca099871951dc05d0006c Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.436607 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.442837 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.446623 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.946587185 +0000 UTC m=+139.220882989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.449900 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.451109 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-5bsgf"] Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.452139 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:14.95212199 +0000 UTC m=+139.226417794 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.458016 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.464809 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-pwx2p"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.468198 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-j84bh" podStartSLOduration=120.463842167 podStartE2EDuration="2m0.463842167s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:14.408790705 +0000 UTC m=+138.683086509" watchObservedRunningTime="2025-12-04 11:45:14.463842167 +0000 UTC m=+138.738137971" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.488426 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" podStartSLOduration=121.488395601 podStartE2EDuration="2m1.488395601s" podCreationTimestamp="2025-12-04 11:43:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:14.431660673 +0000 UTC m=+138.705956477" watchObservedRunningTime="2025-12-04 11:45:14.488395601 +0000 UTC m=+138.762691405" Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.492713 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.504665 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-67g6s"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.551997 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.556312 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.056249828 +0000 UTC m=+139.330545632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.562528 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.564027 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.064008312 +0000 UTC m=+139.338304116 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.598055 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-vbkql"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.604527 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.622250 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx"] Dec 04 11:45:14 crc kubenswrapper[4979]: W1204 11:45:14.663711 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c628ec2_7934_4ec0_b08c_24936a63d8fb.slice/crio-80b528ea45909aae38688564fa8921e68bb584a2ead97750c0b35bd8476ab0e1 WatchSource:0}: Error finding container 80b528ea45909aae38688564fa8921e68bb584a2ead97750c0b35bd8476ab0e1: Status 404 returned error can't find the container with id 80b528ea45909aae38688564fa8921e68bb584a2ead97750c0b35bd8476ab0e1 Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.664324 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.664811 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.164788563 +0000 UTC m=+139.439084377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.765504 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.765990 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.265974084 +0000 UTC m=+139.540269888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.780909 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.796399 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csp56"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.806383 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-dgmfk"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.822110 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-g2jss"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.843222 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.845820 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-h6d8x"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.876185 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.876245 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cc929"] Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.876954 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.376931731 +0000 UTC m=+139.651227535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.877100 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.877545 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.377537587 +0000 UTC m=+139.651833391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.894827 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.896385 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-qzm9p"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.907253 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.915615 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.933160 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc"] Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.978843 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.979086 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.479050877 +0000 UTC m=+139.753346681 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:14 crc kubenswrapper[4979]: I1204 11:45:14.980696 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:14 crc kubenswrapper[4979]: E1204 11:45:14.981321 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.481278085 +0000 UTC m=+139.755573889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.059272 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.071094 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:15 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:15 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:15 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.071150 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.085627 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.086379 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.586184754 +0000 UTC m=+139.860480558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.191432 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.192397 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.692378167 +0000 UTC m=+139.966673971 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.292376 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.292712 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.792688835 +0000 UTC m=+140.066984639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.292824 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.293700 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.793691081 +0000 UTC m=+140.067986885 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.395999 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.396356 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:15.896338911 +0000 UTC m=+140.170634715 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.444365 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" event={"ID":"749499f8-a6c6-403c-9a4a-26242d44b3c5","Type":"ContainerStarted","Data":"7ef7c3351f2e41ef02673b1ceb106dd2b3c9ee85c7c05fb6e9e7ab77e209030c"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.460895 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" event={"ID":"dfb0a0f6-642a-469a-891f-5d7bb9a76b01","Type":"ContainerStarted","Data":"72339d75c14dccd6936d9f61d24165651e885197c1be186c29fdec9383cffb3f"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.475201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" event={"ID":"aa07299a-4ae2-4a48-972e-0b2be5c3db10","Type":"ContainerStarted","Data":"1d2547394ae132460a674e91bda0cde1946c173dfbf263f9ff1b9d7a78c97694"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.475249 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" event={"ID":"aa07299a-4ae2-4a48-972e-0b2be5c3db10","Type":"ContainerStarted","Data":"2624f4d8eb6e9b81c9f977ee84d3736ed9029487540bf28e6068e2416fbd03aa"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.477526 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-x46bj" podStartSLOduration=121.477514908 podStartE2EDuration="2m1.477514908s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.477125848 +0000 UTC m=+139.751421652" watchObservedRunningTime="2025-12-04 11:45:15.477514908 +0000 UTC m=+139.751810712" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.486662 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" event={"ID":"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5","Type":"ContainerStarted","Data":"0b27bb1a0dcfbf4694976fc1dce3b9d39e1a5e8aae66e74c6538846329e7a74a"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.501661 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.503072 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.003053677 +0000 UTC m=+140.277349541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.541999 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" podStartSLOduration=120.541972517 podStartE2EDuration="2m0.541972517s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.541926645 +0000 UTC m=+139.816222449" watchObservedRunningTime="2025-12-04 11:45:15.541972517 +0000 UTC m=+139.816268331" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.560565 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" event={"ID":"37364f76-6a05-44ef-b24d-7fc0cd69145d","Type":"ContainerStarted","Data":"b6fb9f939605ded7c6c50cda5223e1641e351afec63b4088d7f6d812fd0720aa"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.562331 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" event={"ID":"45b72490-12d9-450f-9a00-2d467c68a35b","Type":"ContainerStarted","Data":"37cf967fda811b3f80c358f71b9f8e4c47dfba20b8015b2450595504ba075682"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.564401 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" event={"ID":"5592da75-33af-4003-96eb-542cf6cdc256","Type":"ContainerStarted","Data":"5b1c6dba49fde2ac912475fc2206e7cca643585c6a0b0cbbd116cca2dfef65f6"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.569115 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" event={"ID":"9c53cf1a-a171-429e-b852-27132e72cc90","Type":"ContainerStarted","Data":"4e86a9b31982ebd1d183316adfe3a84ce10c7f759cb6688e01ee3354988855d1"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.569209 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" event={"ID":"9c53cf1a-a171-429e-b852-27132e72cc90","Type":"ContainerStarted","Data":"a3664cbd91a80381980af3dc798d7b4cb121b29b837eb82537e90dffff437854"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.577571 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.594704 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.608269 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-c6fzf" podStartSLOduration=120.608253024 podStartE2EDuration="2m0.608253024s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.606688792 +0000 UTC m=+139.880984596" watchObservedRunningTime="2025-12-04 11:45:15.608253024 +0000 UTC m=+139.882548828" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.609777 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.610969 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.110955264 +0000 UTC m=+140.385251068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.632626 4979 generic.go:334] "Generic (PLEG): container finished" podID="bb732274-6b17-47f6-ace6-485706ecd538" containerID="6ea63990cdaef579e1d69fdff853394f2242baa32dcbf70ecd87311c85d604e4" exitCode=0 Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.632756 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" event={"ID":"bb732274-6b17-47f6-ace6-485706ecd538","Type":"ContainerStarted","Data":"e643e6e86e9195dc9dfd0ba57f18fec1a05f62e2c0c431b92f41e76b4b8e680b"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.632814 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" event={"ID":"bb732274-6b17-47f6-ace6-485706ecd538","Type":"ContainerDied","Data":"6ea63990cdaef579e1d69fdff853394f2242baa32dcbf70ecd87311c85d604e4"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.633948 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.647251 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" event={"ID":"e1cac92c-afb8-4309-9fb8-28cd2f334abd","Type":"ContainerStarted","Data":"ca8e35a0d1d7529ca847aae0c6ecfbf55481b425291082ec2b929e6fc1d63d5e"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.651115 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" event={"ID":"5056ac5a-1bc7-4298-9417-995c16c28ec1","Type":"ContainerStarted","Data":"b881ddd0f875f4c8eced365bf644f0a379297991180bbe2a4cc85c36bfb1f80d"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.662946 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" event={"ID":"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c","Type":"ContainerStarted","Data":"fd924f5219492bdc39b68debc700537b92d064103fef1c2f0f2cbf623694dbeb"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.732328 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.733081 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" podStartSLOduration=121.733026923 podStartE2EDuration="2m1.733026923s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.729460519 +0000 UTC m=+140.003756333" watchObservedRunningTime="2025-12-04 11:45:15.733026923 +0000 UTC m=+140.007322727" Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.733779 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.233761042 +0000 UTC m=+140.508056886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.771638 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" event={"ID":"307823df-a4bd-4d96-a33c-f1684b6a00a6","Type":"ContainerStarted","Data":"516f2fdece56da499e29a4ee6bd85844f4162d314f18e9478fc4798881f42124"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.771683 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" event={"ID":"307823df-a4bd-4d96-a33c-f1684b6a00a6","Type":"ContainerStarted","Data":"26da29cab630c0c8d02f9655a6567bbcb2cffa3c9fb6e581191b7993c7305630"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.771835 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.772912 4979 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-jbsjs container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" start-of-body= Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.772955 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" podUID="307823df-a4bd-4d96-a33c-f1684b6a00a6" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.20:5443/healthz\": dial tcp 10.217.0.20:5443: connect: connection refused" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.797441 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" podStartSLOduration=120.797417521 podStartE2EDuration="2m0.797417521s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.795754716 +0000 UTC m=+140.070050530" watchObservedRunningTime="2025-12-04 11:45:15.797417521 +0000 UTC m=+140.071713325" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.799920 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" event={"ID":"1b73cd49-492e-443e-b199-b7f9b80b066c","Type":"ContainerStarted","Data":"108d3ebc962e22801c48aca7e2eedaf53b26e96252ac332c58ee91f3a0a1ca5a"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.799993 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" event={"ID":"1b73cd49-492e-443e-b199-b7f9b80b066c","Type":"ContainerStarted","Data":"ce07e52f31f10d1946f93ae332ae5c2088d197d36769243735d3be8f3de0fb93"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.804187 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" event={"ID":"8930c76d-0bde-451d-aea5-50323292fb6b","Type":"ContainerStarted","Data":"1905fddae9f4e8d48476659de0257b9e4952754171f59ee963201c1fedfae6dc"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.815114 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" event={"ID":"9c628ec2-7934-4ec0-b08c-24936a63d8fb","Type":"ContainerStarted","Data":"80b528ea45909aae38688564fa8921e68bb584a2ead97750c0b35bd8476ab0e1"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.827242 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c54gv" podStartSLOduration=121.827217261 podStartE2EDuration="2m1.827217261s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.826810441 +0000 UTC m=+140.101106235" watchObservedRunningTime="2025-12-04 11:45:15.827217261 +0000 UTC m=+140.101513065" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.830592 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" event={"ID":"16cbfdaf-1b4c-463e-9767-6879dc885056","Type":"ContainerStarted","Data":"29db5cbe3770f70c2bcb572272301f2146d5c1d641b0fda1af33d476fe6b3faa"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.836867 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.837071 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.337046919 +0000 UTC m=+140.611342733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.837212 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.838253 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.33824106 +0000 UTC m=+140.612536864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.842230 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" event={"ID":"cd25b52b-0498-4bb0-a818-a674ab7b0eee","Type":"ContainerStarted","Data":"031b43ba072a72c9122eb4783af7aef77e17387299a89c8d29392098c9d05d0b"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.842281 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" event={"ID":"cd25b52b-0498-4bb0-a818-a674ab7b0eee","Type":"ContainerStarted","Data":"af25925f8fba0642e5dc9ccbc528cfd9edd426137cd1021f8419ac703c32d846"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.851952 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" event={"ID":"5a9c53b8-81fc-4e05-bb9b-f205aea24e47","Type":"ContainerStarted","Data":"673aac4d24e147be0883875f6d42b0e189ee48bdf0ca0d9e942050c500827487"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.853067 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" event={"ID":"3f367570-d5b9-48b3-a76b-9931605c22cd","Type":"ContainerStarted","Data":"5dc2e1d0b5452c85456eeed608f8d213300ad05f73524b2899390e5d54b8ac98"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.861269 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" event={"ID":"74542077-889e-48b4-9af3-3f0ffde94fc8","Type":"ContainerStarted","Data":"60f71692c2766875c69747358c254b291c8941db5d664e7de014fe40e92e7cd1"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.862955 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" event={"ID":"de406636-3ee1-4b50-b102-8167b8a87963","Type":"ContainerStarted","Data":"5d9bf9d6f5970f124121a33abaa25e24d1b22e5166ecfaba9dee771327625c79"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.867565 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vdmr2" podStartSLOduration=121.867541068 podStartE2EDuration="2m1.867541068s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.866837919 +0000 UTC m=+140.141133733" watchObservedRunningTime="2025-12-04 11:45:15.867541068 +0000 UTC m=+140.141836872" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.874995 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cc929" event={"ID":"0d1cf870-bdba-419c-9534-2f8cdf10e1c7","Type":"ContainerStarted","Data":"0f8e6cf856e7fa9f2fe2d6aed4ff5e77003e0c62c2268f78c5d5ca8362921f29"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.899105 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qwjms" event={"ID":"ee371f5c-b62b-4694-808b-f1f26a2353c5","Type":"ContainerStarted","Data":"9b509a3a4fadb91a9de2ad13201aed0e684e89c7741b16e4fca576a0f2eb11b7"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.899169 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-qwjms" event={"ID":"ee371f5c-b62b-4694-808b-f1f26a2353c5","Type":"ContainerStarted","Data":"ba14dae0483d7119205d47ca1fb8bc6a4c578cb1596a714562fc9385d33aa37e"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.902211 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.903241 4979 patch_prober.go:28] interesting pod/downloads-7954f5f757-qwjms container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.903287 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qwjms" podUID="ee371f5c-b62b-4694-808b-f1f26a2353c5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.906383 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" podStartSLOduration=121.906355694 podStartE2EDuration="2m1.906355694s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.905085392 +0000 UTC m=+140.179381206" watchObservedRunningTime="2025-12-04 11:45:15.906355694 +0000 UTC m=+140.180651498" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.925003 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5wfhh" event={"ID":"82926c75-1f3c-4230-b981-f341f4e17ee2","Type":"ContainerStarted","Data":"2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.938022 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:15 crc kubenswrapper[4979]: E1204 11:45:15.939438 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.439419221 +0000 UTC m=+140.713715025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.942623 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-qwjms" podStartSLOduration=121.942605834 podStartE2EDuration="2m1.942605834s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.941112425 +0000 UTC m=+140.215408249" watchObservedRunningTime="2025-12-04 11:45:15.942605834 +0000 UTC m=+140.216901638" Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.945378 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" event={"ID":"3bc70b2d-20e0-4336-acf6-8606980f9b87","Type":"ContainerStarted","Data":"b201e50a1e96818030b6c0b6ed2f71aba5a3625c6e6b2d8917c2cd4fb8b0fdbc"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.945433 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" event={"ID":"3bc70b2d-20e0-4336-acf6-8606980f9b87","Type":"ContainerStarted","Data":"7f4a422f29401a80ccf7977022f24ac381bd89912f3ca099871951dc05d0006c"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.976575 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" event={"ID":"a27662f3-0c05-4afb-addc-ebcfdf5a6b95","Type":"ContainerStarted","Data":"db082265a08bc26d65cc5fc5a5affb208e31d0f929850af98e4f509b8d6e919b"} Dec 04 11:45:15 crc kubenswrapper[4979]: I1204 11:45:15.977027 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" event={"ID":"a27662f3-0c05-4afb-addc-ebcfdf5a6b95","Type":"ContainerStarted","Data":"c6ed0391105e73ea9fba1411693b2c4e98ec3e97f9bbac799a8099ca23814135"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.015426 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-5wfhh" podStartSLOduration=122.015403782 podStartE2EDuration="2m2.015403782s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:15.976860162 +0000 UTC m=+140.251155966" watchObservedRunningTime="2025-12-04 11:45:16.015403782 +0000 UTC m=+140.289699596" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.029509 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-m55jc" event={"ID":"4a8680c4-28b9-4242-8249-ad3efcd042a0","Type":"ContainerStarted","Data":"9de8d2143c24cb7bbda9e78e851f3c77c6c1e9f732c0a5239b0129a082ac16a1"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.040133 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.040585 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.540567091 +0000 UTC m=+140.814862895 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.052674 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" event={"ID":"e3bd54a5-561b-4207-9c2b-24718a0b473f","Type":"ContainerStarted","Data":"8a03a4945f37137684d5e7971c4d3806c0a6a00a157166b903a6a95d5ab6687c"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.052731 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" event={"ID":"e3bd54a5-561b-4207-9c2b-24718a0b473f","Type":"ContainerStarted","Data":"2c2da2f4878ad18bd46db429f2adc4d60aaa87fe0f930cdeceef3c643d1df970"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.053964 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.054587 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-nzww8" podStartSLOduration=122.054577108 podStartE2EDuration="2m2.054577108s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:16.019449168 +0000 UTC m=+140.293744972" watchObservedRunningTime="2025-12-04 11:45:16.054577108 +0000 UTC m=+140.328872902" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.056561 4979 patch_prober.go:28] interesting pod/console-operator-58897d9998-xmjpc container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.056725 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" podUID="e3bd54a5-561b-4207-9c2b-24718a0b473f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.9:8443/readyz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.057687 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-dkjdk" podStartSLOduration=122.05767208 podStartE2EDuration="2m2.05767208s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:16.055194095 +0000 UTC m=+140.329489899" watchObservedRunningTime="2025-12-04 11:45:16.05767208 +0000 UTC m=+140.331967884" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.070354 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:16 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:16 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:16 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.070938 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.094596 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" event={"ID":"c2582dcb-a8bc-4c67-8e7f-43c1303112d6","Type":"ContainerStarted","Data":"c9408476de98ef8010a637ba3784eb7b07affd38514d139914fb23c57d50f56e"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.096013 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" podStartSLOduration=122.096000964 podStartE2EDuration="2m2.096000964s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:16.094075754 +0000 UTC m=+140.368371558" watchObservedRunningTime="2025-12-04 11:45:16.096000964 +0000 UTC m=+140.370296768" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.121693 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" event={"ID":"9f27cd34-9226-4968-a4e7-83788b460159","Type":"ContainerStarted","Data":"f6f497c4efa75eede492e8cd9e421a6de60cc5cdc88dc6785f0fc18954ee4eec"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.121743 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" event={"ID":"9f27cd34-9226-4968-a4e7-83788b460159","Type":"ContainerStarted","Data":"f5a572ede50317646fc05f2c852f6b82b2614d28232ab7f787cdfc61bcc8ca50"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.125610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" event={"ID":"cc835cc1-b1bd-472d-9839-85427e5a6ad7","Type":"ContainerStarted","Data":"c1a8f6536a07e11ceb3b688d6374ae91c03b78b79bfb8e18c8371a7a76a4389f"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.127247 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" event={"ID":"ca6bec4f-27fb-4c21-a7cc-b4611385dc27","Type":"ContainerStarted","Data":"acbf019560d63da1afe445e04e0fb20ff609eaf5214b1baf0044fafb885d552b"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.128444 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-g2jss" event={"ID":"bf77b7e3-6944-4940-b090-bc5a1d3c2b43","Type":"ContainerStarted","Data":"94897ad28a9749ea462ab3e9fafd156a3d44867ef9a852d0bb9f4c1e89eea009"} Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.140990 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.141237 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.641198038 +0000 UTC m=+140.915493842 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.141752 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.144567 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.644549626 +0000 UTC m=+140.918845430 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.246235 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.253542 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.753504871 +0000 UTC m=+141.027800675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.354972 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.355775 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.85575145 +0000 UTC m=+141.130047254 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.456388 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.456502 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.956485309 +0000 UTC m=+141.230781113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.456732 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.457048 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:16.957041844 +0000 UTC m=+141.231337648 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.559976 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.560992 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.060973377 +0000 UTC m=+141.335269181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.670952 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.671427 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.171410261 +0000 UTC m=+141.445706075 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.773521 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.773888 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.273873426 +0000 UTC m=+141.548169230 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.875121 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.876008 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.375989082 +0000 UTC m=+141.650284886 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.905559 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kwwbv"] Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.906828 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.908727 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.926288 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kwwbv"] Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.990764 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.991737 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-utilities\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.991818 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-catalog-content\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:16 crc kubenswrapper[4979]: I1204 11:45:16.991919 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl5vc\" (UniqueName: \"kubernetes.io/projected/96fc77cc-ab61-462b-9e6a-0877bb125da6-kube-api-access-hl5vc\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:16 crc kubenswrapper[4979]: E1204 11:45:16.998164 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.498113572 +0000 UTC m=+141.772409376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.063379 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:17 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:17 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:17 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.063482 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.093235 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-utilities\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.093334 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-catalog-content\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.093418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl5vc\" (UniqueName: \"kubernetes.io/projected/96fc77cc-ab61-462b-9e6a-0877bb125da6-kube-api-access-hl5vc\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.093465 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.093819 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.593800279 +0000 UTC m=+141.868096083 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.107229 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-utilities\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.107267 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-catalog-content\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.127575 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5k8fs"] Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.130281 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.146712 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.180503 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5k8fs"] Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.183013 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl5vc\" (UniqueName: \"kubernetes.io/projected/96fc77cc-ab61-462b-9e6a-0877bb125da6-kube-api-access-hl5vc\") pod \"community-operators-kwwbv\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.196629 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.196758 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.696732126 +0000 UTC m=+141.971027940 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.196991 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.197032 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-utilities\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.197179 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzmtf\" (UniqueName: \"kubernetes.io/projected/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-kube-api-access-xzmtf\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.197221 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-catalog-content\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.197571 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.697561488 +0000 UTC m=+141.971857292 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.240791 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" event={"ID":"aa07299a-4ae2-4a48-972e-0b2be5c3db10","Type":"ContainerStarted","Data":"705f7affc537aff6861079afd3728e37d0a23663e3756dac02db90bf5b48958e"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.241367 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.296609 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-thnm7" podStartSLOduration=123.296584352 podStartE2EDuration="2m3.296584352s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:17.270324585 +0000 UTC m=+141.544620399" watchObservedRunningTime="2025-12-04 11:45:17.296584352 +0000 UTC m=+141.570880156" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.298097 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.298422 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.79839937 +0000 UTC m=+142.072695174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.298519 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzmtf\" (UniqueName: \"kubernetes.io/projected/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-kube-api-access-xzmtf\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.298581 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-catalog-content\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.298646 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.298674 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-utilities\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.300691 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.800672829 +0000 UTC m=+142.074968633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.310154 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-utilities\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.311617 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j62qj"] Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.312954 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-catalog-content\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.316374 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.318187 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" event={"ID":"5a9c53b8-81fc-4e05-bb9b-f205aea24e47","Type":"ContainerStarted","Data":"df8ff46e75eff99de6b87a8aec62da8654528a71a0d18d1b8d5ec835ae84b4ed"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.321256 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j62qj"] Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.346913 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzmtf\" (UniqueName: \"kubernetes.io/projected/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-kube-api-access-xzmtf\") pod \"certified-operators-5k8fs\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.367793 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" event={"ID":"c2582dcb-a8bc-4c67-8e7f-43c1303112d6","Type":"ContainerStarted","Data":"491ffc005d7eefa9c385d52abfe6560cb8281a086c1b2dba528d387661ab102b"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.375410 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-29vsx" event={"ID":"3f367570-d5b9-48b3-a76b-9931605c22cd","Type":"ContainerStarted","Data":"b681e0828ef37c30ff3fa5b86de6110065102c245e67508e28171e2431c40734"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.400074 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.400414 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xljnj\" (UniqueName: \"kubernetes.io/projected/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-kube-api-access-xljnj\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.400461 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-utilities\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.400479 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-catalog-content\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.400763 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:17.900735162 +0000 UTC m=+142.175030966 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.423485 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" event={"ID":"9f27cd34-9226-4968-a4e7-83788b460159","Type":"ContainerStarted","Data":"49c0f3b9c6d2a1ce4bf2eb900f0017b1b760198576c9eafa52ec49d13ab39e11"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.469487 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" event={"ID":"ca6bec4f-27fb-4c21-a7cc-b4611385dc27","Type":"ContainerStarted","Data":"269dfc190732ecadf973e8dd58e8c2a4bbce9f2cd673da438771091b091b8b53"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.502439 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-5bsgf" podStartSLOduration=123.502413965 podStartE2EDuration="2m3.502413965s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:17.455567698 +0000 UTC m=+141.729863502" watchObservedRunningTime="2025-12-04 11:45:17.502413965 +0000 UTC m=+141.776709769" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.507275 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.507351 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xljnj\" (UniqueName: \"kubernetes.io/projected/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-kube-api-access-xljnj\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.507402 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-utilities\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.507423 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-catalog-content\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.509127 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-catalog-content\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.509533 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.009516771 +0000 UTC m=+142.283812575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.510424 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-utilities\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.511541 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9lv68"] Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.512902 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.533772 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lv68"] Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.535081 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-g2jss" event={"ID":"bf77b7e3-6944-4940-b090-bc5a1d3c2b43","Type":"ContainerStarted","Data":"8b6ffba2d9e63e079ce9a794d5480b3b1fe06d7c42cdae2ee9676034ac628091"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.577062 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xljnj\" (UniqueName: \"kubernetes.io/projected/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-kube-api-access-xljnj\") pod \"community-operators-j62qj\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.604475 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" event={"ID":"cc835cc1-b1bd-472d-9839-85427e5a6ad7","Type":"ContainerStarted","Data":"bf2da3e5a3ef1d8eadbf6f7c32e5cba5da974fb235d827bf84256cf123c65964"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.614399 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.614983 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.114961715 +0000 UTC m=+142.389257519 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.628469 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" event={"ID":"74542077-889e-48b4-9af3-3f0ffde94fc8","Type":"ContainerStarted","Data":"8f8e329d3b92ae3868989856523ea31b185ab402d813a0a2dea67ebff89de8f4"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.631126 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" event={"ID":"16cbfdaf-1b4c-463e-9767-6879dc885056","Type":"ContainerStarted","Data":"6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.631941 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.634533 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.646995 4979 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-csp56 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.647072 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.671923 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.680484 4979 generic.go:334] "Generic (PLEG): container finished" podID="5592da75-33af-4003-96eb-542cf6cdc256" containerID="a4d00fe6d64c073dccacca64d0a4cb154ac5d03086dbc3cdb27a03b75166479e" exitCode=0 Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.680635 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" event={"ID":"5592da75-33af-4003-96eb-542cf6cdc256","Type":"ContainerDied","Data":"a4d00fe6d64c073dccacca64d0a4cb154ac5d03086dbc3cdb27a03b75166479e"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.681476 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-g2jss" podStartSLOduration=8.681451277 podStartE2EDuration="8.681451277s" podCreationTimestamp="2025-12-04 11:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:17.622079192 +0000 UTC m=+141.896374996" watchObservedRunningTime="2025-12-04 11:45:17.681451277 +0000 UTC m=+141.955747081" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.682197 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-tbthq" podStartSLOduration=123.682190367 podStartE2EDuration="2m3.682190367s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:17.654289046 +0000 UTC m=+141.928584860" watchObservedRunningTime="2025-12-04 11:45:17.682190367 +0000 UTC m=+141.956486181" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.742639 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" podStartSLOduration=122.7426173 podStartE2EDuration="2m2.7426173s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:17.697263242 +0000 UTC m=+141.971559056" watchObservedRunningTime="2025-12-04 11:45:17.7426173 +0000 UTC m=+142.016913104" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.746073 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.746168 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-utilities\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.746195 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-catalog-content\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.746349 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kcc5\" (UniqueName: \"kubernetes.io/projected/9ad983ec-486e-4f86-b938-79b2f72639e2-kube-api-access-2kcc5\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.760657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" event={"ID":"37364f76-6a05-44ef-b24d-7fc0cd69145d","Type":"ContainerStarted","Data":"54996db728f9591db03f7dcc362fcfcf27887739f2e52db92e1a453eb273e74e"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.762241 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.764416 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.264394021 +0000 UTC m=+142.538689825 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.807265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" event={"ID":"de406636-3ee1-4b50-b102-8167b8a87963","Type":"ContainerStarted","Data":"71847ab3aecc284cb6b44d3e728c166289aced9bd3d2a931d93c8c8cbdda06fc"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.848962 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.849190 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-utilities\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.849213 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-catalog-content\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.849277 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kcc5\" (UniqueName: \"kubernetes.io/projected/9ad983ec-486e-4f86-b938-79b2f72639e2-kube-api-access-2kcc5\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.850498 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.350480096 +0000 UTC m=+142.624775900 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.851363 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-utilities\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.875614 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-catalog-content\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.892216 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" podStartSLOduration=123.892195389 podStartE2EDuration="2m3.892195389s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:17.797331053 +0000 UTC m=+142.071626887" watchObservedRunningTime="2025-12-04 11:45:17.892195389 +0000 UTC m=+142.166491193" Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.930791 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" event={"ID":"dfb0a0f6-642a-469a-891f-5d7bb9a76b01","Type":"ContainerStarted","Data":"93831fe2151610bae6be5d29e3d7340579c12d54993835c57bf2b925b43bed36"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.930856 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" event={"ID":"dfb0a0f6-642a-469a-891f-5d7bb9a76b01","Type":"ContainerStarted","Data":"a8b260e32623e096ed74efb1250ab090cc71cd28a325171c7ac01929204e140f"} Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.954164 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:17 crc kubenswrapper[4979]: E1204 11:45:17.955993 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.45597876 +0000 UTC m=+142.730274564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:17 crc kubenswrapper[4979]: I1204 11:45:17.985613 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-67g6s" event={"ID":"a2ba54a7-a273-41ce-b7a1-7dc4949a94d5","Type":"ContainerStarted","Data":"982700094ad3876bc5600f1858349b8016beaa12ba53fbb907f678ebccb1d416"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.034773 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kwwbv"] Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.048600 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" event={"ID":"9c628ec2-7934-4ec0-b08c-24936a63d8fb","Type":"ContainerStarted","Data":"e8ebdb4c0442301b0bfcf7bb29874cdd0ad921e4187902a03fbe05d3d988caef"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.057034 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.061773 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.561750002 +0000 UTC m=+142.836045806 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.085234 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kcc5\" (UniqueName: \"kubernetes.io/projected/9ad983ec-486e-4f86-b938-79b2f72639e2-kube-api-access-2kcc5\") pod \"certified-operators-9lv68\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.086078 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:18 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:18 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:18 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.086124 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.088701 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-99f98" podStartSLOduration=124.088676757 podStartE2EDuration="2m4.088676757s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.085961736 +0000 UTC m=+142.360257530" watchObservedRunningTime="2025-12-04 11:45:18.088676757 +0000 UTC m=+142.362972561" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.116663 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cc929" event={"ID":"0d1cf870-bdba-419c-9534-2f8cdf10e1c7","Type":"ContainerStarted","Data":"55c421893e5311f39f9dbb0520c8951dd466401f7abb1173e3fd75607ca09e7e"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.134867 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-dgmfk" podStartSLOduration=123.134847827 podStartE2EDuration="2m3.134847827s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.13383387 +0000 UTC m=+142.408129664" watchObservedRunningTime="2025-12-04 11:45:18.134847827 +0000 UTC m=+142.409143631" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.165313 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.165956 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.665920781 +0000 UTC m=+142.940216585 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.180733 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" event={"ID":"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c","Type":"ContainerStarted","Data":"22d92705c2e6c6d98d99b64ff19ce02bc85cfe5a666054928ea4ec73b506c701"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.180773 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" event={"ID":"3a3ca47d-a6ad-40dc-8d75-d7cca9e6060c","Type":"ContainerStarted","Data":"c54ef0194f617de975f6cf420e6d571d7cf2e72966fe396fe5d193d859f36bb3"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.184703 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.199146 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-64jzx" podStartSLOduration=124.199122762 podStartE2EDuration="2m4.199122762s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.180542135 +0000 UTC m=+142.454837939" watchObservedRunningTime="2025-12-04 11:45:18.199122762 +0000 UTC m=+142.473418566" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.247690 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-q5vs8" podStartSLOduration=124.247669924 podStartE2EDuration="2m4.247669924s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.245664221 +0000 UTC m=+142.519960025" watchObservedRunningTime="2025-12-04 11:45:18.247669924 +0000 UTC m=+142.521965718" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.262321 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" event={"ID":"45b72490-12d9-450f-9a00-2d467c68a35b","Type":"ContainerStarted","Data":"3efc9a5888abf3c6cd89453932ff08f0afca923fd1688adb642de8e184a40ddf"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.266999 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.268756 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.768730756 +0000 UTC m=+143.043026560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.291202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" event={"ID":"5056ac5a-1bc7-4298-9417-995c16c28ec1","Type":"ContainerStarted","Data":"8b201fdda6f56d0c79110919245a41b83c144ddc3c0dab7efb19a389dc3e71d6"} Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.296466 4979 patch_prober.go:28] interesting pod/downloads-7954f5f757-qwjms container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.296536 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qwjms" podUID="ee371f5c-b62b-4694-808b-f1f26a2353c5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.322656 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-jbsjs" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.328821 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-xmjpc" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.371031 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.382834 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.882812154 +0000 UTC m=+143.157108038 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.427482 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-cv8nc" podStartSLOduration=124.427459274 podStartE2EDuration="2m4.427459274s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.324635361 +0000 UTC m=+142.598931175" watchObservedRunningTime="2025-12-04 11:45:18.427459274 +0000 UTC m=+142.701755078" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.466219 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" podStartSLOduration=18.452043289 podStartE2EDuration="18.452043289s" podCreationTimestamp="2025-12-04 11:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.426055208 +0000 UTC m=+142.700351012" watchObservedRunningTime="2025-12-04 11:45:18.452043289 +0000 UTC m=+142.726339093" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.467062 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-vbkql" podStartSLOduration=124.467045492 podStartE2EDuration="2m4.467045492s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:18.389958282 +0000 UTC m=+142.664254076" watchObservedRunningTime="2025-12-04 11:45:18.467045492 +0000 UTC m=+142.741341296" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.471870 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.472235 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:18.972216907 +0000 UTC m=+143.246512711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.494533 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j62qj"] Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.575565 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.575937 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.075922645 +0000 UTC m=+143.350218449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.676445 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.677486 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.177463726 +0000 UTC m=+143.451759540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.796461 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.797408 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.297222063 +0000 UTC m=+143.571517867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.881623 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5k8fs"] Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.901238 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:18 crc kubenswrapper[4979]: E1204 11:45:18.901740 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.401716102 +0000 UTC m=+143.676011906 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.906048 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6zl"] Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.907294 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.926812 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 11:45:18 crc kubenswrapper[4979]: I1204 11:45:18.930527 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6zl"] Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.003580 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-utilities\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.003680 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6ptw\" (UniqueName: \"kubernetes.io/projected/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-kube-api-access-x6ptw\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.003729 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-catalog-content\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.003777 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.004171 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.504152345 +0000 UTC m=+143.778448159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.065561 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:19 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:19 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:19 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.065641 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.106028 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.106263 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.60621496 +0000 UTC m=+143.880510764 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.106595 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-utilities\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.106651 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6ptw\" (UniqueName: \"kubernetes.io/projected/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-kube-api-access-x6ptw\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.106690 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-catalog-content\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.106730 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.107041 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.607025831 +0000 UTC m=+143.881321635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.107875 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-utilities\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.108180 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-catalog-content\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.137717 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6ptw\" (UniqueName: \"kubernetes.io/projected/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-kube-api-access-x6ptw\") pod \"redhat-marketplace-vz6zl\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.164947 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9lv68"] Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.207827 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.208035 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.708006907 +0000 UTC m=+143.982302711 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.208155 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.208592 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.708582872 +0000 UTC m=+143.982878676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.240801 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.298389 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vsl8r"] Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.299661 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.310128 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.310527 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-catalog-content\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.310556 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hghl2\" (UniqueName: \"kubernetes.io/projected/adbe29e3-587a-4bd3-bc13-e9577c1db522-kube-api-access-hghl2\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.310594 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-utilities\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.310700 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.810683198 +0000 UTC m=+144.084978992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.342820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerStarted","Data":"c836e9a071d0217ab866834eef4e1a20570a76cbfdeffee56b97f7d220e51cd3"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.348882 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vsl8r"] Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.353077 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" event={"ID":"ca6bec4f-27fb-4c21-a7cc-b4611385dc27","Type":"ContainerStarted","Data":"6c4cb43491427ec050db12fe27f74fd58cb24dac7267d72f303b99c7316e68ea"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.368512 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerStarted","Data":"3b0b02c3c753ac91c940ca18a5ef0412eda0e59a79156feac491c32fa15cea6d"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.368570 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerStarted","Data":"0a4caca6ea785d2e34ebc82c03d3a1a8ae84db7ee6b6c8dc6d14cfb183e3bf72"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.376157 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-pwx2p" podStartSLOduration=124.376139783 podStartE2EDuration="2m4.376139783s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:19.372115937 +0000 UTC m=+143.646411741" watchObservedRunningTime="2025-12-04 11:45:19.376139783 +0000 UTC m=+143.650435587" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.399245 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" event={"ID":"5592da75-33af-4003-96eb-542cf6cdc256","Type":"ContainerStarted","Data":"43be4c7014a9fa42b2e2341d091c137ad075641355b01e13ee621802047ec977"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.401536 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" event={"ID":"e1cac92c-afb8-4309-9fb8-28cd2f334abd","Type":"ContainerStarted","Data":"c9492dbbb575391998a91c372b8a5f5889f96229cd36b34e340a0031970a9f7a"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.402896 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" event={"ID":"5a9c53b8-81fc-4e05-bb9b-f205aea24e47","Type":"ContainerStarted","Data":"6d7ae6d32154426080333f9c2843e7d41b54d7ab8d4be4e51fba7cc9fbed27b5"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.412325 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.412629 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-catalog-content\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.412657 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hghl2\" (UniqueName: \"kubernetes.io/projected/adbe29e3-587a-4bd3-bc13-e9577c1db522-kube-api-access-hghl2\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.412724 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-utilities\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.414249 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:19.914234831 +0000 UTC m=+144.188530635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.414273 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-utilities\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.414328 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-catalog-content\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.414964 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k8fs" event={"ID":"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea","Type":"ContainerStarted","Data":"ab51031a284b7fd248de011320b6e3f8882a43e51d5c136357e2ec17c4f08f97"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.416182 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cc929" event={"ID":"0d1cf870-bdba-419c-9534-2f8cdf10e1c7","Type":"ContainerStarted","Data":"f6bc289bd1e0797f881693eabdcc78d718e95902b0f26d79cc9cbfda0bea2dd9"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.416674 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cc929" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.417596 4979 generic.go:334] "Generic (PLEG): container finished" podID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerID="013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50" exitCode=0 Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.417632 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwwbv" event={"ID":"96fc77cc-ab61-462b-9e6a-0877bb125da6","Type":"ContainerDied","Data":"013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.417647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwwbv" event={"ID":"96fc77cc-ab61-462b-9e6a-0877bb125da6","Type":"ContainerStarted","Data":"3ffc07c5303641ccfa8b0d4c71c28d1ad12050bc99566477caf003c43d01d396"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.419329 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.432366 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" podStartSLOduration=124.432351176 podStartE2EDuration="2m4.432351176s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:19.431803871 +0000 UTC m=+143.706099675" watchObservedRunningTime="2025-12-04 11:45:19.432351176 +0000 UTC m=+143.706646980" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.439439 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" event={"ID":"8930c76d-0bde-451d-aea5-50323292fb6b","Type":"ContainerStarted","Data":"946c1d06566739c710e7f7dc8dc96f18f9a7bab248ca94b5a5b877286cc8e343"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.446155 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hghl2\" (UniqueName: \"kubernetes.io/projected/adbe29e3-587a-4bd3-bc13-e9577c1db522-kube-api-access-hghl2\") pod \"redhat-marketplace-vsl8r\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.474019 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" event={"ID":"c2582dcb-a8bc-4c67-8e7f-43c1303112d6","Type":"ContainerStarted","Data":"025130291eccb8d45dc5fc63de3471ef934ad2430a94863cb576bb99233c1b3c"} Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.474073 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.487544 4979 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-csp56 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.487599 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.487972 4979 patch_prober.go:28] interesting pod/downloads-7954f5f757-qwjms container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.487988 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qwjms" podUID="ee371f5c-b62b-4694-808b-f1f26a2353c5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.510582 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-h6d8x" podStartSLOduration=125.510559025 podStartE2EDuration="2m5.510559025s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:19.471657256 +0000 UTC m=+143.745953070" watchObservedRunningTime="2025-12-04 11:45:19.510559025 +0000 UTC m=+143.784854829" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.513535 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.514887 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.014870088 +0000 UTC m=+144.289165892 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.515090 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.524594 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.024577972 +0000 UTC m=+144.298873776 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.580171 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-9fmlb" podStartSLOduration=125.580153188 podStartE2EDuration="2m5.580153188s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:19.512893146 +0000 UTC m=+143.787188950" watchObservedRunningTime="2025-12-04 11:45:19.580153188 +0000 UTC m=+143.854448992" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.626008 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.626181 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.626266 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.126245366 +0000 UTC m=+144.400541170 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.627027 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.627537 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.127526909 +0000 UTC m=+144.401822713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.646930 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cc929" podStartSLOduration=10.646898997 podStartE2EDuration="10.646898997s" podCreationTimestamp="2025-12-04 11:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:19.640588092 +0000 UTC m=+143.914883896" watchObservedRunningTime="2025-12-04 11:45:19.646898997 +0000 UTC m=+143.921194801" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.716147 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" podStartSLOduration=124.716118251 podStartE2EDuration="2m4.716118251s" podCreationTimestamp="2025-12-04 11:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:19.703408908 +0000 UTC m=+143.977704722" watchObservedRunningTime="2025-12-04 11:45:19.716118251 +0000 UTC m=+143.990414065" Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.733098 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.733415 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.233381724 +0000 UTC m=+144.507677538 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.733553 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.733962 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.233953708 +0000 UTC m=+144.508249522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.751514 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6zl"] Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.834662 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.834847 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.334814881 +0000 UTC m=+144.609110685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.835340 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.835810 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.335791587 +0000 UTC m=+144.610087391 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.935950 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.936127 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.436092645 +0000 UTC m=+144.710388449 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:19 crc kubenswrapper[4979]: I1204 11:45:19.936337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:19 crc kubenswrapper[4979]: E1204 11:45:19.936901 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.436877175 +0000 UTC m=+144.711172979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.037544 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.038390 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.538369805 +0000 UTC m=+144.812665609 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.067498 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:20 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:20 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:20 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.067563 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.091091 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gqrbf"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.092651 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.100741 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.110716 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gqrbf"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.114935 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vsl8r"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.140209 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.141742 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-utilities\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.141796 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-catalog-content\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.141900 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.141937 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx5c8\" (UniqueName: \"kubernetes.io/projected/bc617bb8-ffc1-48b8-9342-136d54548e2e-kube-api-access-mx5c8\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.142351 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.642333749 +0000 UTC m=+144.916629553 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.143246 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.146188 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.146678 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.162182 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.242750 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.242901 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-utilities\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.243064 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.743029508 +0000 UTC m=+145.017325312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.243188 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-catalog-content\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.243350 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.243445 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx5c8\" (UniqueName: \"kubernetes.io/projected/bc617bb8-ffc1-48b8-9342-136d54548e2e-kube-api-access-mx5c8\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.243538 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.243616 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.243783 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-utilities\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.244039 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.744030453 +0000 UTC m=+145.018326247 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.244230 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-catalog-content\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.262955 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx5c8\" (UniqueName: \"kubernetes.io/projected/bc617bb8-ffc1-48b8-9342-136d54548e2e-kube-api-access-mx5c8\") pod \"redhat-operators-gqrbf\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.344067 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.344213 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.844190878 +0000 UTC m=+145.118486682 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.344339 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.344374 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.344401 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.344501 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.344713 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.844702611 +0000 UTC m=+145.118998415 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.365635 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.412175 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.445357 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.445564 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:20.945523213 +0000 UTC m=+145.219819017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.467672 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.479777 4979 generic.go:334] "Generic (PLEG): container finished" podID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerID="b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66" exitCode=0 Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.480149 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k8fs" event={"ID":"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea","Type":"ContainerDied","Data":"b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66"} Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.483038 4979 generic.go:334] "Generic (PLEG): container finished" podID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerID="64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29" exitCode=0 Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.483126 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerDied","Data":"64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29"} Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.490135 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qsmb5"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.492260 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vsl8r" event={"ID":"adbe29e3-587a-4bd3-bc13-e9577c1db522","Type":"ContainerStarted","Data":"40c2c3ec8770cda7761c654bcf3e5e687f13bf08e7404cfcaf3bf40a6032af91"} Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.492402 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.494055 4979 generic.go:334] "Generic (PLEG): container finished" podID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerID="3b0b02c3c753ac91c940ca18a5ef0412eda0e59a79156feac491c32fa15cea6d" exitCode=0 Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.494127 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerDied","Data":"3b0b02c3c753ac91c940ca18a5ef0412eda0e59a79156feac491c32fa15cea6d"} Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.497592 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6zl" event={"ID":"190bcf1a-d610-48d5-a8b5-d94f2369b8d3","Type":"ContainerStarted","Data":"8040c58669939a1fc6176da42b154811deb95d7c0216c880f326c4af3aca73d9"} Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.501905 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qsmb5"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.547932 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvlcj\" (UniqueName: \"kubernetes.io/projected/c5df29a9-d587-465b-a781-9cb74edda3ea-kube-api-access-zvlcj\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.548062 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-utilities\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.549853 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.550081 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-catalog-content\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.550492 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.050473653 +0000 UTC m=+145.324769467 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.655632 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.655854 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.155820504 +0000 UTC m=+145.430116308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.655891 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-catalog-content\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.655928 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvlcj\" (UniqueName: \"kubernetes.io/projected/c5df29a9-d587-465b-a781-9cb74edda3ea-kube-api-access-zvlcj\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.655961 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-utilities\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.656031 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.656363 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.156348778 +0000 UTC m=+145.430644582 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.656922 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-catalog-content\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.657149 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-utilities\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.683359 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvlcj\" (UniqueName: \"kubernetes.io/projected/c5df29a9-d587-465b-a781-9cb74edda3ea-kube-api-access-zvlcj\") pod \"redhat-operators-qsmb5\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.757782 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.758357 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.25832586 +0000 UTC m=+145.532621664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.758642 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.759010 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.258992877 +0000 UTC m=+145.533288691 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.816076 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.866605 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.866814 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.366778731 +0000 UTC m=+145.641074535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.867334 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.867823 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.367806478 +0000 UTC m=+145.642102282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.918541 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gqrbf"] Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.969347 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:20 crc kubenswrapper[4979]: E1204 11:45:20.970198 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.470180501 +0000 UTC m=+145.744476305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:20 crc kubenswrapper[4979]: I1204 11:45:20.987845 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.064100 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:21 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:21 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:21 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.064646 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.071496 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.071892 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.571877636 +0000 UTC m=+145.846173440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: W1204 11:45:21.107491 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podc67ba5d4_da7b_4f2e_90f0_9b1d8c2469d4.slice/crio-b16792d3ef6ae1d8f38d32ea17a8e7515a603d7db44f0b7286de802ffe27fc66 WatchSource:0}: Error finding container b16792d3ef6ae1d8f38d32ea17a8e7515a603d7db44f0b7286de802ffe27fc66: Status 404 returned error can't find the container with id b16792d3ef6ae1d8f38d32ea17a8e7515a603d7db44f0b7286de802ffe27fc66 Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.175851 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.176563 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.676541349 +0000 UTC m=+145.950837153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.278349 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.278883 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.778857469 +0000 UTC m=+146.053153263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.379919 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.380417 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.88040097 +0000 UTC m=+146.154696774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.380654 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qsmb5"] Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.428977 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.429059 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.437169 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.483654 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.486286 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:21.986267794 +0000 UTC m=+146.260563668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.548644 4979 generic.go:334] "Generic (PLEG): container finished" podID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerID="5679b3fa1b5f7dad8e903a8b1c102fa7bcfab64e4ff690151b54b87fe1e74801" exitCode=0 Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.548710 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6zl" event={"ID":"190bcf1a-d610-48d5-a8b5-d94f2369b8d3","Type":"ContainerDied","Data":"5679b3fa1b5f7dad8e903a8b1c102fa7bcfab64e4ff690151b54b87fe1e74801"} Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.567899 4979 generic.go:334] "Generic (PLEG): container finished" podID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerID="1fa16e871f06b51d967f1716cb27971cd1278f203801188bb2f68c442a5983d0" exitCode=0 Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.568029 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vsl8r" event={"ID":"adbe29e3-587a-4bd3-bc13-e9577c1db522","Type":"ContainerDied","Data":"1fa16e871f06b51d967f1716cb27971cd1278f203801188bb2f68c442a5983d0"} Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.577511 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.578229 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsmb5" event={"ID":"c5df29a9-d587-465b-a781-9cb74edda3ea","Type":"ContainerStarted","Data":"ad40ffcaba5e96fdd95e899bbad778c724ab9cacd1e6af82779c4eb47f1d90ed"} Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.578517 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.582509 4979 patch_prober.go:28] interesting pod/console-f9d7485db-5wfhh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.582719 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-5wfhh" podUID="82926c75-1f3c-4230-b981-f341f4e17ee2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.584910 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.585432 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.085409242 +0000 UTC m=+146.359705036 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.608535 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerID="e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f" exitCode=0 Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.608636 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqrbf" event={"ID":"bc617bb8-ffc1-48b8-9342-136d54548e2e","Type":"ContainerDied","Data":"e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f"} Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.608662 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqrbf" event={"ID":"bc617bb8-ffc1-48b8-9342-136d54548e2e","Type":"ContainerStarted","Data":"dfc6b31300abde26e873e0a57ef5a962c7dce359e5832b9457603ad5f3189b15"} Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.612802 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.612871 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.638492 4979 patch_prober.go:28] interesting pod/downloads-7954f5f757-qwjms container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.638536 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-qwjms" podUID="ee371f5c-b62b-4694-808b-f1f26a2353c5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.638589 4979 patch_prober.go:28] interesting pod/downloads-7954f5f757-qwjms container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.638601 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-qwjms" podUID="ee371f5c-b62b-4694-808b-f1f26a2353c5" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.27:8080/\": dial tcp 10.217.0.27:8080: connect: connection refused" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.650608 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4","Type":"ContainerStarted","Data":"b16792d3ef6ae1d8f38d32ea17a8e7515a603d7db44f0b7286de802ffe27fc66"} Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.669268 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-dkvxc" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.688590 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.688951 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.188934284 +0000 UTC m=+146.463230118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.790262 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.790760 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.290742012 +0000 UTC m=+146.565037816 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.892119 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.900441 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.400419056 +0000 UTC m=+146.674714860 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.924365 4979 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 04 11:45:21 crc kubenswrapper[4979]: I1204 11:45:21.993520 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:21 crc kubenswrapper[4979]: E1204 11:45:21.994026 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.494004658 +0000 UTC m=+146.768300462 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.017810 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.059170 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.066379 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:22 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:22 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:22 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.066452 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.095529 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:22 crc kubenswrapper[4979]: E1204 11:45:22.097506 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.597470359 +0000 UTC m=+146.871766163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.120910 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.197065 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.197410 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:22 crc kubenswrapper[4979]: E1204 11:45:22.197520 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.697482 +0000 UTC m=+146.971777804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.197889 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.198279 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.228091 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.299033 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.299113 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.299177 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:22 crc kubenswrapper[4979]: E1204 11:45:22.301769 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-04 11:45:22.801723811 +0000 UTC m=+147.076019675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-stbh4" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.302585 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.335965 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.340195 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.359768 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.360351 4979 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-04T11:45:21.924772664Z","Handler":null,"Name":""} Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.362907 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.373568 4979 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.373617 4979 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.399845 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.416035 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.502211 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.562901 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.562957 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.629872 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-stbh4\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.632883 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.782501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" event={"ID":"8930c76d-0bde-451d-aea5-50323292fb6b","Type":"ContainerStarted","Data":"a029b4a063f1736329581d5414c03b8e215e4fd7737183ab3083ad21406af913"} Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.784186 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" event={"ID":"8930c76d-0bde-451d-aea5-50323292fb6b","Type":"ContainerStarted","Data":"4cba80b2af687e7e25abf6527885a0bec9887c0f3815d4b00529088ce9705f56"} Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.787497 4979 generic.go:334] "Generic (PLEG): container finished" podID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerID="ac7b05bcae0a48250cbf92660828781c12d35a464aec4bae416fbc0d5e911cf9" exitCode=0 Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.787640 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsmb5" event={"ID":"c5df29a9-d587-465b-a781-9cb74edda3ea","Type":"ContainerDied","Data":"ac7b05bcae0a48250cbf92660828781c12d35a464aec4bae416fbc0d5e911cf9"} Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.813874 4979 generic.go:334] "Generic (PLEG): container finished" podID="c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4" containerID="1bd98b363a10d17a56724ba82b037538d2ed4dbc7b2142115c16c8fcacd4c350" exitCode=0 Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.814217 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4","Type":"ContainerDied","Data":"1bd98b363a10d17a56724ba82b037538d2ed4dbc7b2142115c16c8fcacd4c350"} Dec 04 11:45:22 crc kubenswrapper[4979]: I1204 11:45:22.824064 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-r59vv" Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.074378 4979 patch_prober.go:28] interesting pod/router-default-5444994796-j84bh container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 04 11:45:23 crc kubenswrapper[4979]: [-]has-synced failed: reason withheld Dec 04 11:45:23 crc kubenswrapper[4979]: [+]process-running ok Dec 04 11:45:23 crc kubenswrapper[4979]: healthz check failed Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.074446 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-j84bh" podUID="abc2fd1c-2098-4d1e-b86e-9c4e8be715ca" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 04 11:45:23 crc kubenswrapper[4979]: W1204 11:45:23.256063 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-ca57b13672f44a5287fa7dc6d32101379e18c73de67ae507bfc6f64d6498c938 WatchSource:0}: Error finding container ca57b13672f44a5287fa7dc6d32101379e18c73de67ae507bfc6f64d6498c938: Status 404 returned error can't find the container with id ca57b13672f44a5287fa7dc6d32101379e18c73de67ae507bfc6f64d6498c938 Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.291037 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-stbh4"] Dec 04 11:45:23 crc kubenswrapper[4979]: W1204 11:45:23.381963 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-31ba40c6e76781d1d7d5bc9ac30e52936511eb74ca5dc68e3d6cf63d5792343d WatchSource:0}: Error finding container 31ba40c6e76781d1d7d5bc9ac30e52936511eb74ca5dc68e3d6cf63d5792343d: Status 404 returned error can't find the container with id 31ba40c6e76781d1d7d5bc9ac30e52936511eb74ca5dc68e3d6cf63d5792343d Dec 04 11:45:23 crc kubenswrapper[4979]: W1204 11:45:23.383919 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-2d7552e11b587b43ef03cb3b31ca585d315f7eb1f73cd04b73a08053ef7d56fd WatchSource:0}: Error finding container 2d7552e11b587b43ef03cb3b31ca585d315f7eb1f73cd04b73a08053ef7d56fd: Status 404 returned error can't find the container with id 2d7552e11b587b43ef03cb3b31ca585d315f7eb1f73cd04b73a08053ef7d56fd Dec 04 11:45:23 crc kubenswrapper[4979]: W1204 11:45:23.386670 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac3bc99d_9384_4e8b_b266_92aa7a417150.slice/crio-1df2c3d3710577aa8c81e0476a1cec86af63a02121a9210c00a57f43c600968d WatchSource:0}: Error finding container 1df2c3d3710577aa8c81e0476a1cec86af63a02121a9210c00a57f43c600968d: Status 404 returned error can't find the container with id 1df2c3d3710577aa8c81e0476a1cec86af63a02121a9210c00a57f43c600968d Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.854868 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"9608810a29be9e264731f9db4d5075224a71f907523575ad7ec90998c809fc48"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.855254 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"31ba40c6e76781d1d7d5bc9ac30e52936511eb74ca5dc68e3d6cf63d5792343d"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.858119 4979 generic.go:334] "Generic (PLEG): container finished" podID="5056ac5a-1bc7-4298-9417-995c16c28ec1" containerID="8b201fdda6f56d0c79110919245a41b83c144ddc3c0dab7efb19a389dc3e71d6" exitCode=0 Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.858179 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" event={"ID":"5056ac5a-1bc7-4298-9417-995c16c28ec1","Type":"ContainerDied","Data":"8b201fdda6f56d0c79110919245a41b83c144ddc3c0dab7efb19a389dc3e71d6"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.915840 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" event={"ID":"8930c76d-0bde-451d-aea5-50323292fb6b","Type":"ContainerStarted","Data":"d5f972f57a11021b204080aaff406e3220db298daf03574d085e0575e64d3308"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.942112 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-qzm9p" podStartSLOduration=14.942089563 podStartE2EDuration="14.942089563s" podCreationTimestamp="2025-12-04 11:45:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:23.936318972 +0000 UTC m=+148.210614776" watchObservedRunningTime="2025-12-04 11:45:23.942089563 +0000 UTC m=+148.216385367" Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.947893 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"645302ae72cacde01685f5c27b72174e0b73816d41e427e22610cef72607f80d"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.948064 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"2d7552e11b587b43ef03cb3b31ca585d315f7eb1f73cd04b73a08053ef7d56fd"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.978277 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"4ff9722b6aaad65eef7d429e9625ae5dffe018b14c4f961bd010ad8f856c0754"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.978362 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ca57b13672f44a5287fa7dc6d32101379e18c73de67ae507bfc6f64d6498c938"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.978767 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.984984 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" event={"ID":"ac3bc99d-9384-4e8b-b266-92aa7a417150","Type":"ContainerStarted","Data":"1df2c3d3710577aa8c81e0476a1cec86af63a02121a9210c00a57f43c600968d"} Dec 04 11:45:23 crc kubenswrapper[4979]: I1204 11:45:23.985056 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.031940 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" podStartSLOduration=130.031863816 podStartE2EDuration="2m10.031863816s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:24.029992956 +0000 UTC m=+148.304288770" watchObservedRunningTime="2025-12-04 11:45:24.031863816 +0000 UTC m=+148.306159630" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.069716 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.076519 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-j84bh" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.235812 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.271570 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.272649 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.275900 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.276231 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.285041 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.352350 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695290de-9377-4049-9000-6843a6fdd608-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.352770 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695290de-9377-4049-9000-6843a6fdd608-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.454561 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695290de-9377-4049-9000-6843a6fdd608-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.454603 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695290de-9377-4049-9000-6843a6fdd608-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.454672 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695290de-9377-4049-9000-6843a6fdd608-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.462369 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.493262 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695290de-9377-4049-9000-6843a6fdd608-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.557528 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kube-api-access\") pod \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.557583 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kubelet-dir\") pod \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\" (UID: \"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4\") " Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.557927 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4" (UID: "c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.562359 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4" (UID: "c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.614065 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.660427 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:24 crc kubenswrapper[4979]: I1204 11:45:24.660473 4979 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.020624 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4","Type":"ContainerDied","Data":"b16792d3ef6ae1d8f38d32ea17a8e7515a603d7db44f0b7286de802ffe27fc66"} Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.021149 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b16792d3ef6ae1d8f38d32ea17a8e7515a603d7db44f0b7286de802ffe27fc66" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.021232 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.026461 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" event={"ID":"ac3bc99d-9384-4e8b-b266-92aa7a417150","Type":"ContainerStarted","Data":"413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb"} Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.300570 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.663711 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.800787 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5056ac5a-1bc7-4298-9417-995c16c28ec1-secret-volume\") pod \"5056ac5a-1bc7-4298-9417-995c16c28ec1\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.800843 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume\") pod \"5056ac5a-1bc7-4298-9417-995c16c28ec1\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.800928 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n6ljt\" (UniqueName: \"kubernetes.io/projected/5056ac5a-1bc7-4298-9417-995c16c28ec1-kube-api-access-n6ljt\") pod \"5056ac5a-1bc7-4298-9417-995c16c28ec1\" (UID: \"5056ac5a-1bc7-4298-9417-995c16c28ec1\") " Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.802786 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume" (OuterVolumeSpecName: "config-volume") pod "5056ac5a-1bc7-4298-9417-995c16c28ec1" (UID: "5056ac5a-1bc7-4298-9417-995c16c28ec1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.808508 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5056ac5a-1bc7-4298-9417-995c16c28ec1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5056ac5a-1bc7-4298-9417-995c16c28ec1" (UID: "5056ac5a-1bc7-4298-9417-995c16c28ec1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.809282 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5056ac5a-1bc7-4298-9417-995c16c28ec1-kube-api-access-n6ljt" (OuterVolumeSpecName: "kube-api-access-n6ljt") pod "5056ac5a-1bc7-4298-9417-995c16c28ec1" (UID: "5056ac5a-1bc7-4298-9417-995c16c28ec1"). InnerVolumeSpecName "kube-api-access-n6ljt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.903420 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5056ac5a-1bc7-4298-9417-995c16c28ec1-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.903479 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5056ac5a-1bc7-4298-9417-995c16c28ec1-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:25 crc kubenswrapper[4979]: I1204 11:45:25.903490 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n6ljt\" (UniqueName: \"kubernetes.io/projected/5056ac5a-1bc7-4298-9417-995c16c28ec1-kube-api-access-n6ljt\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:26 crc kubenswrapper[4979]: I1204 11:45:26.072325 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"695290de-9377-4049-9000-6843a6fdd608","Type":"ContainerStarted","Data":"ccabb306fd1fd0d239f85a3eb48a6d9cd5b198c5f1b9e4be6ca23e673d56833a"} Dec 04 11:45:26 crc kubenswrapper[4979]: I1204 11:45:26.108072 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" Dec 04 11:45:26 crc kubenswrapper[4979]: I1204 11:45:26.108264 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc" event={"ID":"5056ac5a-1bc7-4298-9417-995c16c28ec1","Type":"ContainerDied","Data":"b881ddd0f875f4c8eced365bf644f0a379297991180bbe2a4cc85c36bfb1f80d"} Dec 04 11:45:26 crc kubenswrapper[4979]: I1204 11:45:26.108371 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b881ddd0f875f4c8eced365bf644f0a379297991180bbe2a4cc85c36bfb1f80d" Dec 04 11:45:27 crc kubenswrapper[4979]: I1204 11:45:27.492739 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cc929" Dec 04 11:45:28 crc kubenswrapper[4979]: I1204 11:45:28.041139 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:45:28 crc kubenswrapper[4979]: I1204 11:45:28.041222 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:45:28 crc kubenswrapper[4979]: I1204 11:45:28.125784 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"695290de-9377-4049-9000-6843a6fdd608","Type":"ContainerStarted","Data":"5e89f70ac14d4deba84ea251f602da19a68aa64777f3196a685965bbb6f44cc8"} Dec 04 11:45:28 crc kubenswrapper[4979]: I1204 11:45:28.144993 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.144971388 podStartE2EDuration="4.144971388s" podCreationTimestamp="2025-12-04 11:45:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:45:28.144613009 +0000 UTC m=+152.418908813" watchObservedRunningTime="2025-12-04 11:45:28.144971388 +0000 UTC m=+152.419267192" Dec 04 11:45:29 crc kubenswrapper[4979]: I1204 11:45:29.143938 4979 generic.go:334] "Generic (PLEG): container finished" podID="695290de-9377-4049-9000-6843a6fdd608" containerID="5e89f70ac14d4deba84ea251f602da19a68aa64777f3196a685965bbb6f44cc8" exitCode=0 Dec 04 11:45:29 crc kubenswrapper[4979]: I1204 11:45:29.143999 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"695290de-9377-4049-9000-6843a6fdd608","Type":"ContainerDied","Data":"5e89f70ac14d4deba84ea251f602da19a68aa64777f3196a685965bbb6f44cc8"} Dec 04 11:45:31 crc kubenswrapper[4979]: I1204 11:45:31.580710 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:31 crc kubenswrapper[4979]: I1204 11:45:31.584798 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:45:31 crc kubenswrapper[4979]: I1204 11:45:31.613606 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:45:31 crc kubenswrapper[4979]: I1204 11:45:31.642778 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-qwjms" Dec 04 11:45:36 crc kubenswrapper[4979]: I1204 11:45:36.685468 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:36 crc kubenswrapper[4979]: I1204 11:45:36.691662 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e031b763-d441-46af-a2af-ad62bd130a6c-metrics-certs\") pod \"network-metrics-daemon-td9cq\" (UID: \"e031b763-d441-46af-a2af-ad62bd130a6c\") " pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:36 crc kubenswrapper[4979]: I1204 11:45:36.742961 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-td9cq" Dec 04 11:45:40 crc kubenswrapper[4979]: I1204 11:45:40.989678 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.117858 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695290de-9377-4049-9000-6843a6fdd608-kube-api-access\") pod \"695290de-9377-4049-9000-6843a6fdd608\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.117928 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695290de-9377-4049-9000-6843a6fdd608-kubelet-dir\") pod \"695290de-9377-4049-9000-6843a6fdd608\" (UID: \"695290de-9377-4049-9000-6843a6fdd608\") " Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.118123 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/695290de-9377-4049-9000-6843a6fdd608-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "695290de-9377-4049-9000-6843a6fdd608" (UID: "695290de-9377-4049-9000-6843a6fdd608"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.122467 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/695290de-9377-4049-9000-6843a6fdd608-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "695290de-9377-4049-9000-6843a6fdd608" (UID: "695290de-9377-4049-9000-6843a6fdd608"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.219432 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/695290de-9377-4049-9000-6843a6fdd608-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.219463 4979 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/695290de-9377-4049-9000-6843a6fdd608-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.620389 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"695290de-9377-4049-9000-6843a6fdd608","Type":"ContainerDied","Data":"ccabb306fd1fd0d239f85a3eb48a6d9cd5b198c5f1b9e4be6ca23e673d56833a"} Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.620428 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 04 11:45:41 crc kubenswrapper[4979]: I1204 11:45:41.620438 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccabb306fd1fd0d239f85a3eb48a6d9cd5b198c5f1b9e4be6ca23e673d56833a" Dec 04 11:45:42 crc kubenswrapper[4979]: I1204 11:45:42.639206 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:45:52 crc kubenswrapper[4979]: I1204 11:45:52.130224 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" Dec 04 11:45:58 crc kubenswrapper[4979]: I1204 11:45:58.041381 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:45:58 crc kubenswrapper[4979]: I1204 11:45:58.041986 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:45:59 crc kubenswrapper[4979]: E1204 11:45:59.889281 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 11:45:59 crc kubenswrapper[4979]: E1204 11:45:59.889732 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hghl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vsl8r_openshift-marketplace(adbe29e3-587a-4bd3-bc13-e9577c1db522): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:45:59 crc kubenswrapper[4979]: E1204 11:45:59.890953 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vsl8r" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" Dec 04 11:46:00 crc kubenswrapper[4979]: E1204 11:46:00.762039 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vsl8r" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.447948 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 11:46:02 crc kubenswrapper[4979]: E1204 11:46:02.448660 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5056ac5a-1bc7-4298-9417-995c16c28ec1" containerName="collect-profiles" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.448676 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5056ac5a-1bc7-4298-9417-995c16c28ec1" containerName="collect-profiles" Dec 04 11:46:02 crc kubenswrapper[4979]: E1204 11:46:02.448699 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="695290de-9377-4049-9000-6843a6fdd608" containerName="pruner" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.448706 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="695290de-9377-4049-9000-6843a6fdd608" containerName="pruner" Dec 04 11:46:02 crc kubenswrapper[4979]: E1204 11:46:02.448719 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4" containerName="pruner" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.448728 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4" containerName="pruner" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.448850 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c67ba5d4-da7b-4f2e-90f0-9b1d8c2469d4" containerName="pruner" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.448868 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="695290de-9377-4049-9000-6843a6fdd608" containerName="pruner" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.448879 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5056ac5a-1bc7-4298-9417-995c16c28ec1" containerName="collect-profiles" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.449367 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.452758 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.453162 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.457924 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.627836 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.627882 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.679608 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.728970 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.729040 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.729219 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.749472 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:02 crc kubenswrapper[4979]: I1204 11:46:02.772447 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:03 crc kubenswrapper[4979]: E1204 11:46:03.950662 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 11:46:03 crc kubenswrapper[4979]: E1204 11:46:03.951369 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mx5c8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-gqrbf_openshift-marketplace(bc617bb8-ffc1-48b8-9342-136d54548e2e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:46:03 crc kubenswrapper[4979]: E1204 11:46:03.953893 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-gqrbf" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" Dec 04 11:46:04 crc kubenswrapper[4979]: E1204 11:46:04.019566 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 04 11:46:04 crc kubenswrapper[4979]: E1204 11:46:04.019737 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xzmtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5k8fs_openshift-marketplace(c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:46:04 crc kubenswrapper[4979]: E1204 11:46:04.020596 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 04 11:46:04 crc kubenswrapper[4979]: E1204 11:46:04.020724 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-x6ptw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-vz6zl_openshift-marketplace(190bcf1a-d610-48d5-a8b5-d94f2369b8d3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:46:04 crc kubenswrapper[4979]: E1204 11:46:04.021378 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5k8fs" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" Dec 04 11:46:04 crc kubenswrapper[4979]: E1204 11:46:04.022851 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-vz6zl" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.351577 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-gqrbf" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.351624 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-vz6zl" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.351624 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5k8fs" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.445167 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.445881 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xljnj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-j62qj_openshift-marketplace(e98b7196-3e76-423a-9ea2-a39c44bd0a9d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.446956 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-j62qj" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.494928 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.495089 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hl5vc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-kwwbv_openshift-marketplace(96fc77cc-ab61-462b-9e6a-0877bb125da6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.496505 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-kwwbv" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.508160 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.508357 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zvlcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-qsmb5_openshift-marketplace(c5df29a9-d587-465b-a781-9cb74edda3ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.510465 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-qsmb5" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" Dec 04 11:46:05 crc kubenswrapper[4979]: I1204 11:46:05.786086 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerStarted","Data":"a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8"} Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.787473 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-qsmb5" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.788431 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-j62qj" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" Dec 04 11:46:05 crc kubenswrapper[4979]: E1204 11:46:05.788529 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-kwwbv" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" Dec 04 11:46:05 crc kubenswrapper[4979]: I1204 11:46:05.828839 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 04 11:46:05 crc kubenswrapper[4979]: I1204 11:46:05.876921 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-td9cq"] Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.792131 4979 generic.go:334] "Generic (PLEG): container finished" podID="201b4358-9a2e-4fbc-bfbc-9655419cfd16" containerID="a6e9ff181c05b2ff5d62c449db720722ed4fec64b41c91cea6c3980a86488317" exitCode=0 Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.792222 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"201b4358-9a2e-4fbc-bfbc-9655419cfd16","Type":"ContainerDied","Data":"a6e9ff181c05b2ff5d62c449db720722ed4fec64b41c91cea6c3980a86488317"} Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.792608 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"201b4358-9a2e-4fbc-bfbc-9655419cfd16","Type":"ContainerStarted","Data":"5a2a6e324329960662807b8ac7af66a80dc979a8803d4ba9c4c3d92d1ce0f9b5"} Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.794922 4979 generic.go:334] "Generic (PLEG): container finished" podID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerID="a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8" exitCode=0 Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.794977 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerDied","Data":"a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8"} Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.797006 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-td9cq" event={"ID":"e031b763-d441-46af-a2af-ad62bd130a6c","Type":"ContainerStarted","Data":"8db84c768c598bff6bc0ef41d430b49f36668cc1dd34e6090664caac03bb3453"} Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.797041 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-td9cq" event={"ID":"e031b763-d441-46af-a2af-ad62bd130a6c","Type":"ContainerStarted","Data":"222dcfb79f58c971a55962bfad2f5009193ab712b4d659989576bb0ac433d255"} Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.797053 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-td9cq" event={"ID":"e031b763-d441-46af-a2af-ad62bd130a6c","Type":"ContainerStarted","Data":"6eed02c31f1433f2f39d86122d22813490f96dbabfd4493c2640b7923df5034d"} Dec 04 11:46:06 crc kubenswrapper[4979]: I1204 11:46:06.823543 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-td9cq" podStartSLOduration=172.823521649 podStartE2EDuration="2m52.823521649s" podCreationTimestamp="2025-12-04 11:43:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:46:06.820217933 +0000 UTC m=+191.094513757" watchObservedRunningTime="2025-12-04 11:46:06.823521649 +0000 UTC m=+191.097817453" Dec 04 11:46:07 crc kubenswrapper[4979]: I1204 11:46:07.806610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerStarted","Data":"9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e"} Dec 04 11:46:07 crc kubenswrapper[4979]: I1204 11:46:07.827543 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9lv68" podStartSLOduration=4.129553225 podStartE2EDuration="50.827524284s" podCreationTimestamp="2025-12-04 11:45:17 +0000 UTC" firstStartedPulling="2025-12-04 11:45:20.609947712 +0000 UTC m=+144.884243516" lastFinishedPulling="2025-12-04 11:46:07.307918771 +0000 UTC m=+191.582214575" observedRunningTime="2025-12-04 11:46:07.826475176 +0000 UTC m=+192.100770980" watchObservedRunningTime="2025-12-04 11:46:07.827524284 +0000 UTC m=+192.101820088" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.098046 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.112966 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kube-api-access\") pod \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.113026 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kubelet-dir\") pod \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\" (UID: \"201b4358-9a2e-4fbc-bfbc-9655419cfd16\") " Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.113446 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "201b4358-9a2e-4fbc-bfbc-9655419cfd16" (UID: "201b4358-9a2e-4fbc-bfbc-9655419cfd16"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.118351 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "201b4358-9a2e-4fbc-bfbc-9655419cfd16" (UID: "201b4358-9a2e-4fbc-bfbc-9655419cfd16"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.185710 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.186109 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.214468 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.214507 4979 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/201b4358-9a2e-4fbc-bfbc-9655419cfd16-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.446457 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 11:46:08 crc kubenswrapper[4979]: E1204 11:46:08.447053 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201b4358-9a2e-4fbc-bfbc-9655419cfd16" containerName="pruner" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.447071 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="201b4358-9a2e-4fbc-bfbc-9655419cfd16" containerName="pruner" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.447205 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="201b4358-9a2e-4fbc-bfbc-9655419cfd16" containerName="pruner" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.447673 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.467826 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.517826 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kube-api-access\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.517939 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kubelet-dir\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.517968 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-var-lock\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.618544 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kube-api-access\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.618682 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kubelet-dir\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.618732 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-var-lock\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.618832 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-var-lock\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.618888 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kubelet-dir\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.635619 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kube-api-access\") pod \"installer-9-crc\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.768965 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.812920 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"201b4358-9a2e-4fbc-bfbc-9655419cfd16","Type":"ContainerDied","Data":"5a2a6e324329960662807b8ac7af66a80dc979a8803d4ba9c4c3d92d1ce0f9b5"} Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.812975 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a2a6e324329960662807b8ac7af66a80dc979a8803d4ba9c4c3d92d1ce0f9b5" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.813058 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 04 11:46:08 crc kubenswrapper[4979]: I1204 11:46:08.963548 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 04 11:46:09 crc kubenswrapper[4979]: I1204 11:46:09.258825 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-9lv68" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="registry-server" probeResult="failure" output=< Dec 04 11:46:09 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 11:46:09 crc kubenswrapper[4979]: > Dec 04 11:46:09 crc kubenswrapper[4979]: I1204 11:46:09.821439 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8ad8b2c5-93f3-442e-97a3-69fcabe05460","Type":"ContainerStarted","Data":"d2272f3657e31fb36b1819cc6ac8ecb2186bcdf7b219a73567a2cf605bc29595"} Dec 04 11:46:09 crc kubenswrapper[4979]: I1204 11:46:09.821868 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8ad8b2c5-93f3-442e-97a3-69fcabe05460","Type":"ContainerStarted","Data":"62877e87f286824319b02a619df3487425068f2e9a8e32c27c9a46cd507589ae"} Dec 04 11:46:09 crc kubenswrapper[4979]: I1204 11:46:09.840206 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.840185542 podStartE2EDuration="1.840185542s" podCreationTimestamp="2025-12-04 11:46:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:46:09.835485268 +0000 UTC m=+194.109781072" watchObservedRunningTime="2025-12-04 11:46:09.840185542 +0000 UTC m=+194.114481346" Dec 04 11:46:17 crc kubenswrapper[4979]: I1204 11:46:17.869514 4979 generic.go:334] "Generic (PLEG): container finished" podID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerID="5c72358b4d492ccea94bc9d4eb1ee7445caf11e4868402427daf14a6ca20f96f" exitCode=0 Dec 04 11:46:17 crc kubenswrapper[4979]: I1204 11:46:17.869572 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vsl8r" event={"ID":"adbe29e3-587a-4bd3-bc13-e9577c1db522","Type":"ContainerDied","Data":"5c72358b4d492ccea94bc9d4eb1ee7445caf11e4868402427daf14a6ca20f96f"} Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.239623 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.300275 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.882984 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vsl8r" event={"ID":"adbe29e3-587a-4bd3-bc13-e9577c1db522","Type":"ContainerStarted","Data":"9087a244048014f77f6b661c14a7e8acd668b5ab327f2dacbd2dfe46475eaf1f"} Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.886672 4979 generic.go:334] "Generic (PLEG): container finished" podID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerID="b34f4260fd1e5906d0de5a8b843532335391c41fb3a405356b4f744710dc18e5" exitCode=0 Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.886735 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsmb5" event={"ID":"c5df29a9-d587-465b-a781-9cb74edda3ea","Type":"ContainerDied","Data":"b34f4260fd1e5906d0de5a8b843532335391c41fb3a405356b4f744710dc18e5"} Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.893730 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerID="63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979" exitCode=0 Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.893828 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqrbf" event={"ID":"bc617bb8-ffc1-48b8-9342-136d54548e2e","Type":"ContainerDied","Data":"63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979"} Dec 04 11:46:18 crc kubenswrapper[4979]: I1204 11:46:18.910391 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vsl8r" podStartSLOduration=3.13343536 podStartE2EDuration="59.910289037s" podCreationTimestamp="2025-12-04 11:45:19 +0000 UTC" firstStartedPulling="2025-12-04 11:45:21.577540015 +0000 UTC m=+145.851835829" lastFinishedPulling="2025-12-04 11:46:18.354393702 +0000 UTC m=+202.628689506" observedRunningTime="2025-12-04 11:46:18.908790963 +0000 UTC m=+203.183086777" watchObservedRunningTime="2025-12-04 11:46:18.910289037 +0000 UTC m=+203.184584841" Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.626790 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.627425 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.838447 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lv68"] Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.902031 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqrbf" event={"ID":"bc617bb8-ffc1-48b8-9342-136d54548e2e","Type":"ContainerStarted","Data":"6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695"} Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.905413 4979 generic.go:334] "Generic (PLEG): container finished" podID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerID="075f164d5d4b967eec0ec020c86b52080bb68a9f20bc63aec86093e589306678" exitCode=0 Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.905481 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerDied","Data":"075f164d5d4b967eec0ec020c86b52080bb68a9f20bc63aec86093e589306678"} Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.914145 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9lv68" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="registry-server" containerID="cri-o://9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e" gracePeriod=2 Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.914620 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsmb5" event={"ID":"c5df29a9-d587-465b-a781-9cb74edda3ea","Type":"ContainerStarted","Data":"9d0b8a637de8d4dbcfb258ea475ff03ab33c83a0d867503af1096cecd6f8df30"} Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.919109 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gqrbf" podStartSLOduration=2.249452418 podStartE2EDuration="59.919094768s" podCreationTimestamp="2025-12-04 11:45:20 +0000 UTC" firstStartedPulling="2025-12-04 11:45:21.613542059 +0000 UTC m=+145.887837863" lastFinishedPulling="2025-12-04 11:46:19.283184409 +0000 UTC m=+203.557480213" observedRunningTime="2025-12-04 11:46:19.917237115 +0000 UTC m=+204.191532919" watchObservedRunningTime="2025-12-04 11:46:19.919094768 +0000 UTC m=+204.193390572" Dec 04 11:46:19 crc kubenswrapper[4979]: I1204 11:46:19.939068 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qsmb5" podStartSLOduration=3.424584021 podStartE2EDuration="59.939048055s" podCreationTimestamp="2025-12-04 11:45:20 +0000 UTC" firstStartedPulling="2025-12-04 11:45:22.795013096 +0000 UTC m=+147.069308900" lastFinishedPulling="2025-12-04 11:46:19.30947713 +0000 UTC m=+203.583772934" observedRunningTime="2025-12-04 11:46:19.932042352 +0000 UTC m=+204.206338156" watchObservedRunningTime="2025-12-04 11:46:19.939048055 +0000 UTC m=+204.213343859" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.386966 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.414465 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.414534 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.580172 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-utilities\") pod \"9ad983ec-486e-4f86-b938-79b2f72639e2\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.580287 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kcc5\" (UniqueName: \"kubernetes.io/projected/9ad983ec-486e-4f86-b938-79b2f72639e2-kube-api-access-2kcc5\") pod \"9ad983ec-486e-4f86-b938-79b2f72639e2\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.580337 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-catalog-content\") pod \"9ad983ec-486e-4f86-b938-79b2f72639e2\" (UID: \"9ad983ec-486e-4f86-b938-79b2f72639e2\") " Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.580960 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-utilities" (OuterVolumeSpecName: "utilities") pod "9ad983ec-486e-4f86-b938-79b2f72639e2" (UID: "9ad983ec-486e-4f86-b938-79b2f72639e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.587506 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ad983ec-486e-4f86-b938-79b2f72639e2-kube-api-access-2kcc5" (OuterVolumeSpecName: "kube-api-access-2kcc5") pod "9ad983ec-486e-4f86-b938-79b2f72639e2" (UID: "9ad983ec-486e-4f86-b938-79b2f72639e2"). InnerVolumeSpecName "kube-api-access-2kcc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.632278 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9ad983ec-486e-4f86-b938-79b2f72639e2" (UID: "9ad983ec-486e-4f86-b938-79b2f72639e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.665495 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-vsl8r" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="registry-server" probeResult="failure" output=< Dec 04 11:46:20 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 11:46:20 crc kubenswrapper[4979]: > Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.682010 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.682046 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kcc5\" (UniqueName: \"kubernetes.io/projected/9ad983ec-486e-4f86-b938-79b2f72639e2-kube-api-access-2kcc5\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.682059 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9ad983ec-486e-4f86-b938-79b2f72639e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.817035 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.817123 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.921579 4979 generic.go:334] "Generic (PLEG): container finished" podID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerID="d6b9a4d4b768af8b19aa043d5837ffc8f1bd8282e78d3ad96d50db1d1d465b00" exitCode=0 Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.921647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6zl" event={"ID":"190bcf1a-d610-48d5-a8b5-d94f2369b8d3","Type":"ContainerDied","Data":"d6b9a4d4b768af8b19aa043d5837ffc8f1bd8282e78d3ad96d50db1d1d465b00"} Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.924249 4979 generic.go:334] "Generic (PLEG): container finished" podID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerID="9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e" exitCode=0 Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.924531 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerDied","Data":"9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e"} Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.924551 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9lv68" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.924578 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9lv68" event={"ID":"9ad983ec-486e-4f86-b938-79b2f72639e2","Type":"ContainerDied","Data":"c836e9a071d0217ab866834eef4e1a20570a76cbfdeffee56b97f7d220e51cd3"} Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.924590 4979 scope.go:117] "RemoveContainer" containerID="9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.926471 4979 generic.go:334] "Generic (PLEG): container finished" podID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerID="f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe" exitCode=0 Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.927043 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwwbv" event={"ID":"96fc77cc-ab61-462b-9e6a-0877bb125da6","Type":"ContainerDied","Data":"f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe"} Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.950827 4979 scope.go:117] "RemoveContainer" containerID="a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8" Dec 04 11:46:20 crc kubenswrapper[4979]: I1204 11:46:20.974635 4979 scope.go:117] "RemoveContainer" containerID="64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.029984 4979 scope.go:117] "RemoveContainer" containerID="9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e" Dec 04 11:46:21 crc kubenswrapper[4979]: E1204 11:46:21.031855 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e\": container with ID starting with 9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e not found: ID does not exist" containerID="9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.031898 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e"} err="failed to get container status \"9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e\": rpc error: code = NotFound desc = could not find container \"9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e\": container with ID starting with 9f11f832eda2493c98e0b5abc1a2710dccba7ee968df040a2a4edfebb07d0a4e not found: ID does not exist" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.031944 4979 scope.go:117] "RemoveContainer" containerID="a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8" Dec 04 11:46:21 crc kubenswrapper[4979]: E1204 11:46:21.033139 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8\": container with ID starting with a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8 not found: ID does not exist" containerID="a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.033172 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8"} err="failed to get container status \"a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8\": rpc error: code = NotFound desc = could not find container \"a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8\": container with ID starting with a2c0662774ba8066b9da43e4d2c8b5e0a500ae7b2587cf01e9d95a2210bc14c8 not found: ID does not exist" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.033194 4979 scope.go:117] "RemoveContainer" containerID="64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29" Dec 04 11:46:21 crc kubenswrapper[4979]: E1204 11:46:21.033633 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29\": container with ID starting with 64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29 not found: ID does not exist" containerID="64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.033669 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29"} err="failed to get container status \"64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29\": rpc error: code = NotFound desc = could not find container \"64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29\": container with ID starting with 64d170705d31d4e7ad8e1c7d2f8c3020a621b6248e1e750ae8bf7f7750757a29 not found: ID does not exist" Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.036130 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9lv68"] Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.042458 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9lv68"] Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.457891 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gqrbf" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="registry-server" probeResult="failure" output=< Dec 04 11:46:21 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 11:46:21 crc kubenswrapper[4979]: > Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.854759 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qsmb5" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="registry-server" probeResult="failure" output=< Dec 04 11:46:21 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 11:46:21 crc kubenswrapper[4979]: > Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.938780 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwwbv" event={"ID":"96fc77cc-ab61-462b-9e6a-0877bb125da6","Type":"ContainerStarted","Data":"0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419"} Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.945095 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerStarted","Data":"258cd176ae876207ecffb946adefb079358bf370aa1d55f00499b7f6bafd8bc4"} Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.954108 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6zl" event={"ID":"190bcf1a-d610-48d5-a8b5-d94f2369b8d3","Type":"ContainerStarted","Data":"74e33fafe09603ba2cf538b1a6ecbbc353545446a71e08f9475176e2693a27ae"} Dec 04 11:46:21 crc kubenswrapper[4979]: I1204 11:46:21.997082 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kwwbv" podStartSLOduration=3.9670896239999998 podStartE2EDuration="1m5.997054537s" podCreationTimestamp="2025-12-04 11:45:16 +0000 UTC" firstStartedPulling="2025-12-04 11:45:19.419070437 +0000 UTC m=+143.693366241" lastFinishedPulling="2025-12-04 11:46:21.44903534 +0000 UTC m=+205.723331154" observedRunningTime="2025-12-04 11:46:21.976331238 +0000 UTC m=+206.250627062" watchObservedRunningTime="2025-12-04 11:46:21.997054537 +0000 UTC m=+206.271350341" Dec 04 11:46:22 crc kubenswrapper[4979]: I1204 11:46:22.021164 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j62qj" podStartSLOduration=4.806469902 podStartE2EDuration="1m5.021144094s" podCreationTimestamp="2025-12-04 11:45:17 +0000 UTC" firstStartedPulling="2025-12-04 11:45:20.610033724 +0000 UTC m=+144.884329528" lastFinishedPulling="2025-12-04 11:46:20.824707906 +0000 UTC m=+205.099003720" observedRunningTime="2025-12-04 11:46:21.998663354 +0000 UTC m=+206.272959158" watchObservedRunningTime="2025-12-04 11:46:22.021144094 +0000 UTC m=+206.295439898" Dec 04 11:46:22 crc kubenswrapper[4979]: I1204 11:46:22.022593 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vz6zl" podStartSLOduration=4.061894435 podStartE2EDuration="1m4.022588485s" podCreationTimestamp="2025-12-04 11:45:18 +0000 UTC" firstStartedPulling="2025-12-04 11:45:21.550079695 +0000 UTC m=+145.824375499" lastFinishedPulling="2025-12-04 11:46:21.510773755 +0000 UTC m=+205.785069549" observedRunningTime="2025-12-04 11:46:22.021820734 +0000 UTC m=+206.296116538" watchObservedRunningTime="2025-12-04 11:46:22.022588485 +0000 UTC m=+206.296884279" Dec 04 11:46:22 crc kubenswrapper[4979]: I1204 11:46:22.207276 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" path="/var/lib/kubelet/pods/9ad983ec-486e-4f86-b938-79b2f72639e2/volumes" Dec 04 11:46:22 crc kubenswrapper[4979]: I1204 11:46:22.960162 4979 generic.go:334] "Generic (PLEG): container finished" podID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerID="878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4" exitCode=0 Dec 04 11:46:22 crc kubenswrapper[4979]: I1204 11:46:22.960176 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k8fs" event={"ID":"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea","Type":"ContainerDied","Data":"878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4"} Dec 04 11:46:26 crc kubenswrapper[4979]: I1204 11:46:26.982024 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k8fs" event={"ID":"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea","Type":"ContainerStarted","Data":"1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a"} Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.003908 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5k8fs" podStartSLOduration=4.739442194 podStartE2EDuration="1m10.003893572s" podCreationTimestamp="2025-12-04 11:45:17 +0000 UTC" firstStartedPulling="2025-12-04 11:45:20.605912246 +0000 UTC m=+144.880208050" lastFinishedPulling="2025-12-04 11:46:25.870363624 +0000 UTC m=+210.144659428" observedRunningTime="2025-12-04 11:46:27.001787981 +0000 UTC m=+211.276083805" watchObservedRunningTime="2025-12-04 11:46:27.003893572 +0000 UTC m=+211.278189366" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.242178 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.242241 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.286678 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.635040 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.635144 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.673052 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.673090 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.678436 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:46:27 crc kubenswrapper[4979]: I1204 11:46:27.723777 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.020924 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.030910 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.040418 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.040478 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.040525 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.041103 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.041162 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67" gracePeriod=600 Dec 04 11:46:28 crc kubenswrapper[4979]: E1204 11:46:28.124526 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9b48d97_5715_467e_b044_4812d6dfea63.slice/crio-c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67.scope\": RecentStats: unable to find data in memory cache]" Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.992936 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67" exitCode=0 Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.993098 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67"} Dec 04 11:46:28 crc kubenswrapper[4979]: I1204 11:46:28.993151 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"e61fe1ef8585786eb9e6a5ae2992ed20ce3863f21b72946cbe3aff92d1a5bf00"} Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.241709 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.241760 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.287200 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.433102 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j62qj"] Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.664873 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.707769 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:46:29 crc kubenswrapper[4979]: I1204 11:46:29.998649 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j62qj" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="registry-server" containerID="cri-o://258cd176ae876207ecffb946adefb079358bf370aa1d55f00499b7f6bafd8bc4" gracePeriod=2 Dec 04 11:46:30 crc kubenswrapper[4979]: I1204 11:46:30.033183 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:46:30 crc kubenswrapper[4979]: I1204 11:46:30.465415 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:46:30 crc kubenswrapper[4979]: I1204 11:46:30.502515 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:46:30 crc kubenswrapper[4979]: I1204 11:46:30.855255 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:46:30 crc kubenswrapper[4979]: I1204 11:46:30.914851 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:46:31 crc kubenswrapper[4979]: I1204 11:46:31.835102 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vsl8r"] Dec 04 11:46:31 crc kubenswrapper[4979]: I1204 11:46:31.835389 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vsl8r" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="registry-server" containerID="cri-o://9087a244048014f77f6b661c14a7e8acd668b5ab327f2dacbd2dfe46475eaf1f" gracePeriod=2 Dec 04 11:46:32 crc kubenswrapper[4979]: I1204 11:46:32.217449 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-4v569"] Dec 04 11:46:33 crc kubenswrapper[4979]: I1204 11:46:33.016031 4979 generic.go:334] "Generic (PLEG): container finished" podID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerID="9087a244048014f77f6b661c14a7e8acd668b5ab327f2dacbd2dfe46475eaf1f" exitCode=0 Dec 04 11:46:33 crc kubenswrapper[4979]: I1204 11:46:33.016072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vsl8r" event={"ID":"adbe29e3-587a-4bd3-bc13-e9577c1db522","Type":"ContainerDied","Data":"9087a244048014f77f6b661c14a7e8acd668b5ab327f2dacbd2dfe46475eaf1f"} Dec 04 11:46:33 crc kubenswrapper[4979]: I1204 11:46:33.019055 4979 generic.go:334] "Generic (PLEG): container finished" podID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerID="258cd176ae876207ecffb946adefb079358bf370aa1d55f00499b7f6bafd8bc4" exitCode=0 Dec 04 11:46:33 crc kubenswrapper[4979]: I1204 11:46:33.019097 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerDied","Data":"258cd176ae876207ecffb946adefb079358bf370aa1d55f00499b7f6bafd8bc4"} Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.237599 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qsmb5"] Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.238356 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qsmb5" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="registry-server" containerID="cri-o://9d0b8a637de8d4dbcfb258ea475ff03ab33c83a0d867503af1096cecd6f8df30" gracePeriod=2 Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.264957 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.359911 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-utilities\") pod \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.359981 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xljnj\" (UniqueName: \"kubernetes.io/projected/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-kube-api-access-xljnj\") pod \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.360047 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-catalog-content\") pod \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\" (UID: \"e98b7196-3e76-423a-9ea2-a39c44bd0a9d\") " Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.361092 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-utilities" (OuterVolumeSpecName: "utilities") pod "e98b7196-3e76-423a-9ea2-a39c44bd0a9d" (UID: "e98b7196-3e76-423a-9ea2-a39c44bd0a9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.368148 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-kube-api-access-xljnj" (OuterVolumeSpecName: "kube-api-access-xljnj") pod "e98b7196-3e76-423a-9ea2-a39c44bd0a9d" (UID: "e98b7196-3e76-423a-9ea2-a39c44bd0a9d"). InnerVolumeSpecName "kube-api-access-xljnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.414916 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e98b7196-3e76-423a-9ea2-a39c44bd0a9d" (UID: "e98b7196-3e76-423a-9ea2-a39c44bd0a9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.422578 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.464523 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-utilities\") pod \"adbe29e3-587a-4bd3-bc13-e9577c1db522\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.464926 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-catalog-content\") pod \"adbe29e3-587a-4bd3-bc13-e9577c1db522\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.464992 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hghl2\" (UniqueName: \"kubernetes.io/projected/adbe29e3-587a-4bd3-bc13-e9577c1db522-kube-api-access-hghl2\") pod \"adbe29e3-587a-4bd3-bc13-e9577c1db522\" (UID: \"adbe29e3-587a-4bd3-bc13-e9577c1db522\") " Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.465329 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.465353 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.465365 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xljnj\" (UniqueName: \"kubernetes.io/projected/e98b7196-3e76-423a-9ea2-a39c44bd0a9d-kube-api-access-xljnj\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.465504 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-utilities" (OuterVolumeSpecName: "utilities") pod "adbe29e3-587a-4bd3-bc13-e9577c1db522" (UID: "adbe29e3-587a-4bd3-bc13-e9577c1db522"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.476625 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adbe29e3-587a-4bd3-bc13-e9577c1db522-kube-api-access-hghl2" (OuterVolumeSpecName: "kube-api-access-hghl2") pod "adbe29e3-587a-4bd3-bc13-e9577c1db522" (UID: "adbe29e3-587a-4bd3-bc13-e9577c1db522"). InnerVolumeSpecName "kube-api-access-hghl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.491491 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adbe29e3-587a-4bd3-bc13-e9577c1db522" (UID: "adbe29e3-587a-4bd3-bc13-e9577c1db522"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.567036 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.567111 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hghl2\" (UniqueName: \"kubernetes.io/projected/adbe29e3-587a-4bd3-bc13-e9577c1db522-kube-api-access-hghl2\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:34 crc kubenswrapper[4979]: I1204 11:46:34.567131 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adbe29e3-587a-4bd3-bc13-e9577c1db522-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.032539 4979 generic.go:334] "Generic (PLEG): container finished" podID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerID="9d0b8a637de8d4dbcfb258ea475ff03ab33c83a0d867503af1096cecd6f8df30" exitCode=0 Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.032592 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsmb5" event={"ID":"c5df29a9-d587-465b-a781-9cb74edda3ea","Type":"ContainerDied","Data":"9d0b8a637de8d4dbcfb258ea475ff03ab33c83a0d867503af1096cecd6f8df30"} Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.034579 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vsl8r" event={"ID":"adbe29e3-587a-4bd3-bc13-e9577c1db522","Type":"ContainerDied","Data":"40c2c3ec8770cda7761c654bcf3e5e687f13bf08e7404cfcaf3bf40a6032af91"} Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.034597 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vsl8r" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.034615 4979 scope.go:117] "RemoveContainer" containerID="9087a244048014f77f6b661c14a7e8acd668b5ab327f2dacbd2dfe46475eaf1f" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.037478 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j62qj" event={"ID":"e98b7196-3e76-423a-9ea2-a39c44bd0a9d","Type":"ContainerDied","Data":"0a4caca6ea785d2e34ebc82c03d3a1a8ae84db7ee6b6c8dc6d14cfb183e3bf72"} Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.037532 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j62qj" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.053518 4979 scope.go:117] "RemoveContainer" containerID="5c72358b4d492ccea94bc9d4eb1ee7445caf11e4868402427daf14a6ca20f96f" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.069132 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vsl8r"] Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.075261 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.076605 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vsl8r"] Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.084088 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j62qj"] Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.087095 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j62qj"] Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.087620 4979 scope.go:117] "RemoveContainer" containerID="1fa16e871f06b51d967f1716cb27971cd1278f203801188bb2f68c442a5983d0" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.102592 4979 scope.go:117] "RemoveContainer" containerID="258cd176ae876207ecffb946adefb079358bf370aa1d55f00499b7f6bafd8bc4" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.119318 4979 scope.go:117] "RemoveContainer" containerID="075f164d5d4b967eec0ec020c86b52080bb68a9f20bc63aec86093e589306678" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.138730 4979 scope.go:117] "RemoveContainer" containerID="3b0b02c3c753ac91c940ca18a5ef0412eda0e59a79156feac491c32fa15cea6d" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.174509 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-utilities\") pod \"c5df29a9-d587-465b-a781-9cb74edda3ea\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.174613 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvlcj\" (UniqueName: \"kubernetes.io/projected/c5df29a9-d587-465b-a781-9cb74edda3ea-kube-api-access-zvlcj\") pod \"c5df29a9-d587-465b-a781-9cb74edda3ea\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.174666 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-catalog-content\") pod \"c5df29a9-d587-465b-a781-9cb74edda3ea\" (UID: \"c5df29a9-d587-465b-a781-9cb74edda3ea\") " Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.175377 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-utilities" (OuterVolumeSpecName: "utilities") pod "c5df29a9-d587-465b-a781-9cb74edda3ea" (UID: "c5df29a9-d587-465b-a781-9cb74edda3ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.178846 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5df29a9-d587-465b-a781-9cb74edda3ea-kube-api-access-zvlcj" (OuterVolumeSpecName: "kube-api-access-zvlcj") pod "c5df29a9-d587-465b-a781-9cb74edda3ea" (UID: "c5df29a9-d587-465b-a781-9cb74edda3ea"). InnerVolumeSpecName "kube-api-access-zvlcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.276153 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.277276 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvlcj\" (UniqueName: \"kubernetes.io/projected/c5df29a9-d587-465b-a781-9cb74edda3ea-kube-api-access-zvlcj\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.301701 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c5df29a9-d587-465b-a781-9cb74edda3ea" (UID: "c5df29a9-d587-465b-a781-9cb74edda3ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:46:35 crc kubenswrapper[4979]: I1204 11:46:35.378316 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c5df29a9-d587-465b-a781-9cb74edda3ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.046157 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsmb5" event={"ID":"c5df29a9-d587-465b-a781-9cb74edda3ea","Type":"ContainerDied","Data":"ad40ffcaba5e96fdd95e899bbad778c724ab9cacd1e6af82779c4eb47f1d90ed"} Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.046536 4979 scope.go:117] "RemoveContainer" containerID="9d0b8a637de8d4dbcfb258ea475ff03ab33c83a0d867503af1096cecd6f8df30" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.046171 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsmb5" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.073652 4979 scope.go:117] "RemoveContainer" containerID="b34f4260fd1e5906d0de5a8b843532335391c41fb3a405356b4f744710dc18e5" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.094374 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qsmb5"] Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.094576 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qsmb5"] Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.104599 4979 scope.go:117] "RemoveContainer" containerID="ac7b05bcae0a48250cbf92660828781c12d35a464aec4bae416fbc0d5e911cf9" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.205701 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" path="/var/lib/kubelet/pods/adbe29e3-587a-4bd3-bc13-e9577c1db522/volumes" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.206573 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" path="/var/lib/kubelet/pods/c5df29a9-d587-465b-a781-9cb74edda3ea/volumes" Dec 04 11:46:36 crc kubenswrapper[4979]: I1204 11:46:36.207145 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" path="/var/lib/kubelet/pods/e98b7196-3e76-423a-9ea2-a39c44bd0a9d/volumes" Dec 04 11:46:37 crc kubenswrapper[4979]: I1204 11:46:37.678709 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.884537 4979 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885411 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885430 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885446 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885455 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885467 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885473 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885484 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885491 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885501 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885509 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885519 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885529 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885541 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885549 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885560 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885566 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="extract-content" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885577 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885583 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885593 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885602 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885616 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885625 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.885635 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885641 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="extract-utilities" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885785 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ad983ec-486e-4f86-b938-79b2f72639e2" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885805 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e98b7196-3e76-423a-9ea2-a39c44bd0a9d" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885822 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5df29a9-d587-465b-a781-9cb74edda3ea" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.885831 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="adbe29e3-587a-4bd3-bc13-e9577c1db522" containerName="registry-server" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.886359 4979 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.886536 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.887022 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd" gracePeriod=15 Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.887090 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749" gracePeriod=15 Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.887153 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f" gracePeriod=15 Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.887315 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef" gracePeriod=15 Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.887331 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058" gracePeriod=15 Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888207 4979 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.888738 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888764 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.888798 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888808 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.888826 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888833 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.888847 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888854 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.888866 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888872 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.888887 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.888895 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889045 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889080 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889090 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889101 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889110 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889120 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.889240 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 11:46:46 crc kubenswrapper[4979]: I1204 11:46:46.889249 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 04 11:46:46 crc kubenswrapper[4979]: E1204 11:46:46.923471 4979 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.74:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.024926 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025366 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025465 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025569 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025626 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025792 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025844 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.025871 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126717 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126775 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126796 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126809 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126854 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126871 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126833 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126911 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126949 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126974 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.126992 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.127023 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.127043 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.127042 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.127095 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.127140 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.225132 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:47 crc kubenswrapper[4979]: W1204 11:46:47.249344 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-2ac5c4d96cf70ca1170ba642449e1a6aacab09efe43b00be9eabc1c929ccdd32 WatchSource:0}: Error finding container 2ac5c4d96cf70ca1170ba642449e1a6aacab09efe43b00be9eabc1c929ccdd32: Status 404 returned error can't find the container with id 2ac5c4d96cf70ca1170ba642449e1a6aacab09efe43b00be9eabc1c929ccdd32 Dec 04 11:46:47 crc kubenswrapper[4979]: E1204 11:46:47.255236 4979 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.74:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e00a3beaf4a56 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 11:46:47.254641238 +0000 UTC m=+231.528937042,LastTimestamp:2025-12-04 11:46:47.254641238 +0000 UTC m=+231.528937042,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.285241 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.287445 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.288373 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749" exitCode=0 Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.288408 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef" exitCode=0 Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.288421 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f" exitCode=0 Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.288431 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058" exitCode=2 Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.288481 4979 scope.go:117] "RemoveContainer" containerID="1f1513ee70cf656e0aee71f6129f9e120f0d20137df456e3c4c5bbffc9808415" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.295581 4979 generic.go:334] "Generic (PLEG): container finished" podID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" containerID="d2272f3657e31fb36b1819cc6ac8ecb2186bcdf7b219a73567a2cf605bc29595" exitCode=0 Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.295686 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8ad8b2c5-93f3-442e-97a3-69fcabe05460","Type":"ContainerDied","Data":"d2272f3657e31fb36b1819cc6ac8ecb2186bcdf7b219a73567a2cf605bc29595"} Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.296596 4979 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.297157 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:47 crc kubenswrapper[4979]: I1204 11:46:47.304092 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"2ac5c4d96cf70ca1170ba642449e1a6aacab09efe43b00be9eabc1c929ccdd32"} Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.313592 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.317631 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71"} Dec 04 11:46:48 crc kubenswrapper[4979]: E1204 11:46:48.318178 4979 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.74:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.318214 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.554353 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.554931 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.747901 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-var-lock\") pod \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.748088 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-var-lock" (OuterVolumeSpecName: "var-lock") pod "8ad8b2c5-93f3-442e-97a3-69fcabe05460" (UID: "8ad8b2c5-93f3-442e-97a3-69fcabe05460"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.748453 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kube-api-access\") pod \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.748540 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kubelet-dir\") pod \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\" (UID: \"8ad8b2c5-93f3-442e-97a3-69fcabe05460\") " Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.748605 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8ad8b2c5-93f3-442e-97a3-69fcabe05460" (UID: "8ad8b2c5-93f3-442e-97a3-69fcabe05460"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.748826 4979 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.748855 4979 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/8ad8b2c5-93f3-442e-97a3-69fcabe05460-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.755849 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8ad8b2c5-93f3-442e-97a3-69fcabe05460" (UID: "8ad8b2c5-93f3-442e-97a3-69fcabe05460"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:48 crc kubenswrapper[4979]: I1204 11:46:48.849793 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8ad8b2c5-93f3-442e-97a3-69fcabe05460-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.247330 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.249542 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.250151 4979 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.250537 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.326106 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.327111 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd" exitCode=0 Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.327203 4979 scope.go:117] "RemoveContainer" containerID="b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.327268 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.330107 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.331454 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"8ad8b2c5-93f3-442e-97a3-69fcabe05460","Type":"ContainerDied","Data":"62877e87f286824319b02a619df3487425068f2e9a8e32c27c9a46cd507589ae"} Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.331508 4979 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.74:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.331525 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62877e87f286824319b02a619df3487425068f2e9a8e32c27c9a46cd507589ae" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.344284 4979 scope.go:117] "RemoveContainer" containerID="1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.354798 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.354846 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.354883 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.354903 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.354997 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.355087 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.355104 4979 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.355498 4979 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.355527 4979 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.355516 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.355539 4979 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.357837 4979 scope.go:117] "RemoveContainer" containerID="b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.370027 4979 scope.go:117] "RemoveContainer" containerID="03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.384962 4979 scope.go:117] "RemoveContainer" containerID="1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.403806 4979 scope.go:117] "RemoveContainer" containerID="9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.419234 4979 scope.go:117] "RemoveContainer" containerID="b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749" Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.419952 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\": container with ID starting with b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749 not found: ID does not exist" containerID="b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.420035 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749"} err="failed to get container status \"b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\": rpc error: code = NotFound desc = could not find container \"b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749\": container with ID starting with b9d17c327d777cc542de72f0b6fbf5855cd515536be4ea5794f9696e56169749 not found: ID does not exist" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.420069 4979 scope.go:117] "RemoveContainer" containerID="1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef" Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.420684 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\": container with ID starting with 1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef not found: ID does not exist" containerID="1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.420809 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef"} err="failed to get container status \"1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\": rpc error: code = NotFound desc = could not find container \"1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef\": container with ID starting with 1c8f21903cea0d869189bd8ee46ac36eee49b82098b092ad164b8ac0cc1d09ef not found: ID does not exist" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.420950 4979 scope.go:117] "RemoveContainer" containerID="b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f" Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.421564 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\": container with ID starting with b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f not found: ID does not exist" containerID="b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.421598 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f"} err="failed to get container status \"b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\": rpc error: code = NotFound desc = could not find container \"b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f\": container with ID starting with b1fe393c64b23a09780d8332c8e4caa2285cf8c644009b43761e3f2f2dc3a82f not found: ID does not exist" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.421619 4979 scope.go:117] "RemoveContainer" containerID="03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058" Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.421957 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\": container with ID starting with 03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058 not found: ID does not exist" containerID="03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.422056 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058"} err="failed to get container status \"03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\": rpc error: code = NotFound desc = could not find container \"03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058\": container with ID starting with 03002af243fd1c20b1d29f4fd5700051df11808d3d072c2813a1fd2bc1b37058 not found: ID does not exist" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.422153 4979 scope.go:117] "RemoveContainer" containerID="1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd" Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.422490 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\": container with ID starting with 1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd not found: ID does not exist" containerID="1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.422562 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd"} err="failed to get container status \"1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\": rpc error: code = NotFound desc = could not find container \"1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd\": container with ID starting with 1310ef49a5d6909f916de09cf00a744e7d8f5084e28ba501fdb94d81dcd8bafd not found: ID does not exist" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.422586 4979 scope.go:117] "RemoveContainer" containerID="9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713" Dec 04 11:46:49 crc kubenswrapper[4979]: E1204 11:46:49.422972 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\": container with ID starting with 9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713 not found: ID does not exist" containerID="9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.423000 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713"} err="failed to get container status \"9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\": rpc error: code = NotFound desc = could not find container \"9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713\": container with ID starting with 9300b26322fac01c578cd5f2453f99d62a66a196d1ae57870bde87336b11d713 not found: ID does not exist" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.644391 4979 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:49 crc kubenswrapper[4979]: I1204 11:46:49.644930 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: I1204 11:46:50.206802 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.356090 4979 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.74:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e00a3beaf4a56 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-04 11:46:47.254641238 +0000 UTC m=+231.528937042,LastTimestamp:2025-12-04 11:46:47.254641238 +0000 UTC m=+231.528937042,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.647922 4979 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.648549 4979 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.648937 4979 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.649457 4979 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.649802 4979 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: I1204 11:46:50.649837 4979 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.650112 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="200ms" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.775834 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:46:50Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:46:50Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:46:50Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-04T11:46:50Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.776240 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.776476 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.776761 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.777259 4979 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.777334 4979 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 04 11:46:50 crc kubenswrapper[4979]: E1204 11:46:50.851745 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="400ms" Dec 04 11:46:51 crc kubenswrapper[4979]: E1204 11:46:51.252908 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="800ms" Dec 04 11:46:52 crc kubenswrapper[4979]: E1204 11:46:52.054029 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="1.6s" Dec 04 11:46:53 crc kubenswrapper[4979]: E1204 11:46:53.655129 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="3.2s" Dec 04 11:46:56 crc kubenswrapper[4979]: I1204 11:46:56.201836 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:56 crc kubenswrapper[4979]: E1204 11:46:56.856701 4979 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.74:6443: connect: connection refused" interval="6.4s" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.240030 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" podUID="e31c1aff-c394-476a-888e-c0e877f59717" containerName="oauth-openshift" containerID="cri-o://03c276d786e6a1d3eb0c401bf51a63a1cfc2f59c29318e8e705fff1be0737f64" gracePeriod=15 Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.376167 4979 generic.go:334] "Generic (PLEG): container finished" podID="e31c1aff-c394-476a-888e-c0e877f59717" containerID="03c276d786e6a1d3eb0c401bf51a63a1cfc2f59c29318e8e705fff1be0737f64" exitCode=0 Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.376319 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" event={"ID":"e31c1aff-c394-476a-888e-c0e877f59717","Type":"ContainerDied","Data":"03c276d786e6a1d3eb0c401bf51a63a1cfc2f59c29318e8e705fff1be0737f64"} Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.592536 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.592987 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.593245 4979 status_manager.go:851] "Failed to get status for pod" podUID="e31c1aff-c394-476a-888e-c0e877f59717" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4v569\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668066 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-router-certs\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668125 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-session\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668165 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-trusted-ca-bundle\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668191 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-idp-0-file-data\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668213 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-cliconfig\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668238 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-login\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668259 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-service-ca\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668286 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e31c1aff-c394-476a-888e-c0e877f59717-audit-dir\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668333 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-ocp-branding-template\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668361 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-error\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668398 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-424kk\" (UniqueName: \"kubernetes.io/projected/e31c1aff-c394-476a-888e-c0e877f59717-kube-api-access-424kk\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668440 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-serving-cert\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668469 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-audit-policies\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668496 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-provider-selection\") pod \"e31c1aff-c394-476a-888e-c0e877f59717\" (UID: \"e31c1aff-c394-476a-888e-c0e877f59717\") " Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.668915 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e31c1aff-c394-476a-888e-c0e877f59717-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.669330 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.669898 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.670051 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.670195 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.675452 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.675748 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.685963 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.686172 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.686528 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.686995 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31c1aff-c394-476a-888e-c0e877f59717-kube-api-access-424kk" (OuterVolumeSpecName: "kube-api-access-424kk") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "kube-api-access-424kk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.687825 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.688128 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.688313 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e31c1aff-c394-476a-888e-c0e877f59717" (UID: "e31c1aff-c394-476a-888e-c0e877f59717"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770292 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770667 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770693 4979 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e31c1aff-c394-476a-888e-c0e877f59717-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770712 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770729 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770746 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-424kk\" (UniqueName: \"kubernetes.io/projected/e31c1aff-c394-476a-888e-c0e877f59717-kube-api-access-424kk\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770762 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770778 4979 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770795 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770814 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770830 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770845 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770861 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:57 crc kubenswrapper[4979]: I1204 11:46:57.770875 4979 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e31c1aff-c394-476a-888e-c0e877f59717-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.198772 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.199645 4979 status_manager.go:851] "Failed to get status for pod" podUID="e31c1aff-c394-476a-888e-c0e877f59717" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4v569\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.200084 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.214383 4979 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.214423 4979 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:46:58 crc kubenswrapper[4979]: E1204 11:46:58.214831 4979 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.215360 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:58 crc kubenswrapper[4979]: W1204 11:46:58.233263 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-bc942da6ed3268de8abf46f3a20e233fb3b9e02ab6112492005307bb9fcdd36e WatchSource:0}: Error finding container bc942da6ed3268de8abf46f3a20e233fb3b9e02ab6112492005307bb9fcdd36e: Status 404 returned error can't find the container with id bc942da6ed3268de8abf46f3a20e233fb3b9e02ab6112492005307bb9fcdd36e Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.383222 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.383248 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" event={"ID":"e31c1aff-c394-476a-888e-c0e877f59717","Type":"ContainerDied","Data":"84dcf7239b8b0cf85127fe433b4b7779d7e7226140a1ee3e82e0fbc21806918f"} Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.383324 4979 scope.go:117] "RemoveContainer" containerID="03c276d786e6a1d3eb0c401bf51a63a1cfc2f59c29318e8e705fff1be0737f64" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.383949 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.384449 4979 status_manager.go:851] "Failed to get status for pod" podUID="e31c1aff-c394-476a-888e-c0e877f59717" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4v569\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.386034 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bc942da6ed3268de8abf46f3a20e233fb3b9e02ab6112492005307bb9fcdd36e"} Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.388113 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:58 crc kubenswrapper[4979]: I1204 11:46:58.388576 4979 status_manager.go:851] "Failed to get status for pod" podUID="e31c1aff-c394-476a-888e-c0e877f59717" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4v569\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:59 crc kubenswrapper[4979]: I1204 11:46:59.392941 4979 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="963eb3dc1aae41debc9e622f10b0f9d3a86df64e5645f480651efb2b2ca976a6" exitCode=0 Dec 04 11:46:59 crc kubenswrapper[4979]: I1204 11:46:59.393051 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"963eb3dc1aae41debc9e622f10b0f9d3a86df64e5645f480651efb2b2ca976a6"} Dec 04 11:46:59 crc kubenswrapper[4979]: I1204 11:46:59.393415 4979 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:46:59 crc kubenswrapper[4979]: I1204 11:46:59.393446 4979 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:46:59 crc kubenswrapper[4979]: E1204 11:46:59.393893 4979 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:46:59 crc kubenswrapper[4979]: I1204 11:46:59.394101 4979 status_manager.go:851] "Failed to get status for pod" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:46:59 crc kubenswrapper[4979]: I1204 11:46:59.394618 4979 status_manager.go:851] "Failed to get status for pod" podUID="e31c1aff-c394-476a-888e-c0e877f59717" pod="openshift-authentication/oauth-openshift-558db77b4-4v569" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-4v569\": dial tcp 38.102.83.74:6443: connect: connection refused" Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.402681 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.402929 4979 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3" exitCode=1 Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.402968 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3"} Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.403371 4979 scope.go:117] "RemoveContainer" containerID="35013018bcccb5292be35195b971574128d6d5e46ebe8791198dded720b578e3" Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.408451 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"40e17266c0e31368825cbc87f8a8c90bf7de2f816ca648354e45f321c76fbe06"} Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.408488 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2adc4d4751cade25089f470fdcf8a57041c0157e47743270acf56fd34b89bbc9"} Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.408499 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5d27d031dfa460dc46219e6b02253caa30f6744037240436051452e49e1e340c"} Dec 04 11:47:00 crc kubenswrapper[4979]: I1204 11:47:00.408508 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"daa5859d67c71585076f314b733ff7a4d878862f5e9709b071479d7a7d69be31"} Dec 04 11:47:01 crc kubenswrapper[4979]: I1204 11:47:01.417880 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 04 11:47:01 crc kubenswrapper[4979]: I1204 11:47:01.418341 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e1c06764e0328c914461c81e635936055d2818c0c6cf067994256511c953a3dd"} Dec 04 11:47:01 crc kubenswrapper[4979]: I1204 11:47:01.422111 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"f12ccc3e08203ee356600d33d1a48b4f809fc50ac9cecd931d7798dd80b7e781"} Dec 04 11:47:01 crc kubenswrapper[4979]: I1204 11:47:01.422334 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:01 crc kubenswrapper[4979]: I1204 11:47:01.422524 4979 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:47:01 crc kubenswrapper[4979]: I1204 11:47:01.422568 4979 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:47:03 crc kubenswrapper[4979]: I1204 11:47:03.216290 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:03 crc kubenswrapper[4979]: I1204 11:47:03.216700 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:03 crc kubenswrapper[4979]: I1204 11:47:03.222325 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:05 crc kubenswrapper[4979]: I1204 11:47:05.572522 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:47:06 crc kubenswrapper[4979]: I1204 11:47:06.430917 4979 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:06 crc kubenswrapper[4979]: I1204 11:47:06.450603 4979 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:47:06 crc kubenswrapper[4979]: I1204 11:47:06.450642 4979 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:47:06 crc kubenswrapper[4979]: I1204 11:47:06.458736 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:06 crc kubenswrapper[4979]: I1204 11:47:06.462453 4979 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b3a186ca-56fe-422a-96ca-b18530875fae" Dec 04 11:47:07 crc kubenswrapper[4979]: I1204 11:47:07.454413 4979 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:47:07 crc kubenswrapper[4979]: I1204 11:47:07.454744 4979 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="8378641c-2a6c-4f25-a047-f928917e0df5" Dec 04 11:47:10 crc kubenswrapper[4979]: I1204 11:47:10.206696 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:47:10 crc kubenswrapper[4979]: I1204 11:47:10.206803 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:47:10 crc kubenswrapper[4979]: I1204 11:47:10.211925 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 04 11:47:12 crc kubenswrapper[4979]: I1204 11:47:12.509857 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 04 11:47:15 crc kubenswrapper[4979]: I1204 11:47:15.343671 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 04 11:47:16 crc kubenswrapper[4979]: I1204 11:47:16.218504 4979 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b3a186ca-56fe-422a-96ca-b18530875fae" Dec 04 11:47:16 crc kubenswrapper[4979]: I1204 11:47:16.549052 4979 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 04 11:47:17 crc kubenswrapper[4979]: I1204 11:47:17.062341 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 04 11:47:17 crc kubenswrapper[4979]: I1204 11:47:17.343548 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 04 11:47:17 crc kubenswrapper[4979]: I1204 11:47:17.471437 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 04 11:47:17 crc kubenswrapper[4979]: I1204 11:47:17.748841 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 11:47:17 crc kubenswrapper[4979]: I1204 11:47:17.869802 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 04 11:47:17 crc kubenswrapper[4979]: I1204 11:47:17.968270 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 04 11:47:18 crc kubenswrapper[4979]: I1204 11:47:18.091857 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 04 11:47:18 crc kubenswrapper[4979]: I1204 11:47:18.160426 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 04 11:47:18 crc kubenswrapper[4979]: I1204 11:47:18.606078 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 11:47:18 crc kubenswrapper[4979]: I1204 11:47:18.634928 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 04 11:47:18 crc kubenswrapper[4979]: I1204 11:47:18.721077 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 04 11:47:18 crc kubenswrapper[4979]: I1204 11:47:18.868741 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 04 11:47:19 crc kubenswrapper[4979]: I1204 11:47:19.004176 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 04 11:47:19 crc kubenswrapper[4979]: I1204 11:47:19.220347 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 04 11:47:19 crc kubenswrapper[4979]: I1204 11:47:19.439433 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 04 11:47:19 crc kubenswrapper[4979]: I1204 11:47:19.676448 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 04 11:47:19 crc kubenswrapper[4979]: I1204 11:47:19.709682 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 04 11:47:19 crc kubenswrapper[4979]: I1204 11:47:19.879025 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.015130 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.070415 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.203062 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.354758 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.361174 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.432445 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.853580 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.888185 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.913239 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.982206 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 04 11:47:20 crc kubenswrapper[4979]: I1204 11:47:20.985605 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.071992 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.135673 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.192633 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.316785 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.555285 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.615818 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.697556 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.699229 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.728926 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.746527 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.849018 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.958114 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 04 11:47:21 crc kubenswrapper[4979]: I1204 11:47:21.982245 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.124152 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.254885 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.261583 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.306120 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.356782 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.410177 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.433653 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.466437 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.586883 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.613981 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.693600 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.728594 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.761367 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.826336 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 04 11:47:22 crc kubenswrapper[4979]: I1204 11:47:22.861967 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.130232 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.175540 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.211568 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.221888 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.240248 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.247432 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.279746 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.362696 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.486272 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.558187 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.561239 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.677070 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.747498 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.766928 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.797049 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.821839 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.827714 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.985072 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 04 11:47:23 crc kubenswrapper[4979]: I1204 11:47:23.985495 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.068923 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.088038 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.224723 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.262130 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.297190 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.325237 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.338773 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.356637 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.468845 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.526393 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.598449 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.655756 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.690492 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.750409 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.765407 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.842430 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.855625 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.916973 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.944058 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 04 11:47:24 crc kubenswrapper[4979]: I1204 11:47:24.975266 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.095750 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.152578 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.189226 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.209099 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.324168 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.329983 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.343423 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.405661 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.407245 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.525671 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.674350 4979 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.678793 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-4v569"] Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.678848 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.682957 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 04 11:47:25 crc kubenswrapper[4979]: I1204 11:47:25.697358 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=19.697341557 podStartE2EDuration="19.697341557s" podCreationTimestamp="2025-12-04 11:47:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:47:25.694380681 +0000 UTC m=+269.968676485" watchObservedRunningTime="2025-12-04 11:47:25.697341557 +0000 UTC m=+269.971637361" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.047273 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.069534 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.071938 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.157173 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.157549 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.206017 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31c1aff-c394-476a-888e-c0e877f59717" path="/var/lib/kubelet/pods/e31c1aff-c394-476a-888e-c0e877f59717/volumes" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.220918 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.253516 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.292174 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.332190 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.349094 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.489102 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.519270 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.541394 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.603331 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.656750 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.679222 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.751765 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.767951 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.800680 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.803269 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.810389 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.833125 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 04 11:47:26 crc kubenswrapper[4979]: I1204 11:47:26.991319 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.008254 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.109179 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.173667 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.290769 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.354771 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.482281 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.547350 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.550472 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.566832 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.720957 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.810379 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.852848 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.907200 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.955196 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 04 11:47:27 crc kubenswrapper[4979]: I1204 11:47:27.973052 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.043184 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.071715 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.129477 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.131465 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.154687 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.178268 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.250001 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.271473 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.291760 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.357535 4979 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.435510 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.452121 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.605613 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.664987 4979 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.665260 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71" gracePeriod=5 Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.711618 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.770493 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.884240 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 04 11:47:28 crc kubenswrapper[4979]: I1204 11:47:28.921534 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.240993 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.242226 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.242521 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.277233 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.285279 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.314156 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.324714 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.338639 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.438819 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.450599 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.560357 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.573020 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.699939 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.755038 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.891568 4979 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.930051 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 04 11:47:29 crc kubenswrapper[4979]: I1204 11:47:29.952641 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.035001 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.071834 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.219142 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.249641 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.249698 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.311495 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.317572 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.354803 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.383598 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.453012 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.535889 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.567006 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.663535 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.671244 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.712204 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.729744 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.757379 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.765658 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 04 11:47:30 crc kubenswrapper[4979]: I1204 11:47:30.783623 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.043258 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.044337 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.394741 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.396411 4979 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.436592 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.578721 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.864884 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.869766 4979 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.906581 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.918693 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25"] Dec 04 11:47:31 crc kubenswrapper[4979]: E1204 11:47:31.918868 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.918879 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 11:47:31 crc kubenswrapper[4979]: E1204 11:47:31.918887 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31c1aff-c394-476a-888e-c0e877f59717" containerName="oauth-openshift" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.918895 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31c1aff-c394-476a-888e-c0e877f59717" containerName="oauth-openshift" Dec 04 11:47:31 crc kubenswrapper[4979]: E1204 11:47:31.918906 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" containerName="installer" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.918913 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" containerName="installer" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.918995 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31c1aff-c394-476a-888e-c0e877f59717" containerName="oauth-openshift" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.919003 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ad8b2c5-93f3-442e-97a3-69fcabe05460" containerName="installer" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.919011 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.919363 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.924206 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.925883 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.926618 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.926823 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.927166 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.927265 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.928604 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.930804 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.931208 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.931610 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.933210 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.934390 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.949546 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.950857 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.953317 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25"] Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.954317 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 04 11:47:31 crc kubenswrapper[4979]: I1204 11:47:31.982235 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083583 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083638 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-router-certs\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083740 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083761 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083782 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjlzd\" (UniqueName: \"kubernetes.io/projected/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-kube-api-access-hjlzd\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083799 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-login\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083818 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-session\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083910 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-audit-policies\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083943 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-audit-dir\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083965 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-error\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.083984 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.084004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-service-ca\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.084043 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185076 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-audit-policies\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185134 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-audit-dir\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185160 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-error\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185207 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-service-ca\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185229 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185251 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185277 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-router-certs\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185315 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185317 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-audit-dir\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185338 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.185945 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-audit-policies\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.186819 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.186897 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.186902 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.186891 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-service-ca\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.187259 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjlzd\" (UniqueName: \"kubernetes.io/projected/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-kube-api-access-hjlzd\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.187330 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-login\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.187372 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-session\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.190874 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-session\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.190897 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-error\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.190872 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-router-certs\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.191261 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.191712 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.191825 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-login\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.192274 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.202527 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.204717 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjlzd\" (UniqueName: \"kubernetes.io/projected/a7ffeffc-10c1-45cf-b724-bc3dab114a9f-kube-api-access-hjlzd\") pod \"oauth-openshift-5fff7d8cf9-2sk25\" (UID: \"a7ffeffc-10c1-45cf-b724-bc3dab114a9f\") " pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.249861 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.266559 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.427484 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.434634 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25"] Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.547193 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.594407 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" event={"ID":"a7ffeffc-10c1-45cf-b724-bc3dab114a9f","Type":"ContainerStarted","Data":"136243972b281a9e3094ca58a98f73e611471d3a66837c2d03dee4424a853fec"} Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.673529 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.726932 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.901886 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.915778 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 04 11:47:32 crc kubenswrapper[4979]: I1204 11:47:32.984785 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.032197 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.032813 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.238906 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.283052 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.450744 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.551489 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.600977 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5fff7d8cf9-2sk25_a7ffeffc-10c1-45cf-b724-bc3dab114a9f/oauth-openshift/0.log" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.601037 4979 generic.go:334] "Generic (PLEG): container finished" podID="a7ffeffc-10c1-45cf-b724-bc3dab114a9f" containerID="296e50d51346db9c1ddc3eca09198e57c461761b65d92c35d21549edc6de532e" exitCode=255 Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.601068 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" event={"ID":"a7ffeffc-10c1-45cf-b724-bc3dab114a9f","Type":"ContainerDied","Data":"296e50d51346db9c1ddc3eca09198e57c461761b65d92c35d21549edc6de532e"} Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.601484 4979 scope.go:117] "RemoveContainer" containerID="296e50d51346db9c1ddc3eca09198e57c461761b65d92c35d21549edc6de532e" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.788158 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.858983 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.863429 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.907652 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.932294 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 04 11:47:33 crc kubenswrapper[4979]: I1204 11:47:33.988360 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.052509 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.234672 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.234755 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414239 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414316 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414353 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414403 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414454 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414476 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414488 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414509 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414539 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414689 4979 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414701 4979 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414710 4979 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.414720 4979 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.424176 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.515786 4979 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.531468 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.607627 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.607693 4979 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71" exitCode=137 Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.607776 4979 scope.go:117] "RemoveContainer" containerID="58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.607804 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.611718 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-5fff7d8cf9-2sk25_a7ffeffc-10c1-45cf-b724-bc3dab114a9f/oauth-openshift/0.log" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.611780 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" event={"ID":"a7ffeffc-10c1-45cf-b724-bc3dab114a9f","Type":"ContainerStarted","Data":"3fafaeb4cf5a0bedc8966c334f9e5a759b14fcef84bd9a6b8322479d8141121e"} Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.612695 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.621937 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.625905 4979 scope.go:117] "RemoveContainer" containerID="58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71" Dec 04 11:47:34 crc kubenswrapper[4979]: E1204 11:47:34.626426 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71\": container with ID starting with 58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71 not found: ID does not exist" containerID="58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.626519 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71"} err="failed to get container status \"58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71\": rpc error: code = NotFound desc = could not find container \"58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71\": container with ID starting with 58da2356ac2a68aa2ad800eb00d14718e4f2f1aeb798f36445d37af368248f71 not found: ID does not exist" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.642740 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-5fff7d8cf9-2sk25" podStartSLOduration=62.642720773 podStartE2EDuration="1m2.642720773s" podCreationTimestamp="2025-12-04 11:46:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:47:34.641258271 +0000 UTC m=+278.915554075" watchObservedRunningTime="2025-12-04 11:47:34.642720773 +0000 UTC m=+278.917016587" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.697417 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 04 11:47:34 crc kubenswrapper[4979]: I1204 11:47:34.784962 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 04 11:47:35 crc kubenswrapper[4979]: I1204 11:47:35.136942 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 04 11:47:35 crc kubenswrapper[4979]: I1204 11:47:35.278074 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 04 11:47:36 crc kubenswrapper[4979]: I1204 11:47:36.206406 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 04 11:47:40 crc kubenswrapper[4979]: I1204 11:47:40.617222 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-vz6zl" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="registry-server" probeResult="failure" output=< Dec 04 11:47:40 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 11:47:40 crc kubenswrapper[4979]: > Dec 04 11:47:40 crc kubenswrapper[4979]: I1204 11:47:40.619542 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-vz6zl" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="registry-server" probeResult="failure" output=< Dec 04 11:47:40 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 11:47:40 crc kubenswrapper[4979]: > Dec 04 11:47:50 crc kubenswrapper[4979]: I1204 11:47:50.700949 4979 generic.go:334] "Generic (PLEG): container finished" podID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerID="6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804" exitCode=0 Dec 04 11:47:50 crc kubenswrapper[4979]: I1204 11:47:50.701042 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" event={"ID":"16cbfdaf-1b4c-463e-9767-6879dc885056","Type":"ContainerDied","Data":"6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804"} Dec 04 11:47:50 crc kubenswrapper[4979]: I1204 11:47:50.701960 4979 scope.go:117] "RemoveContainer" containerID="6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804" Dec 04 11:47:51 crc kubenswrapper[4979]: I1204 11:47:51.709046 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" event={"ID":"16cbfdaf-1b4c-463e-9767-6879dc885056","Type":"ContainerStarted","Data":"d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961"} Dec 04 11:47:51 crc kubenswrapper[4979]: I1204 11:47:51.710270 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:47:51 crc kubenswrapper[4979]: I1204 11:47:51.712354 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:47:56 crc kubenswrapper[4979]: I1204 11:47:56.067982 4979 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 04 11:48:10 crc kubenswrapper[4979]: I1204 11:48:10.864684 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhnlh"] Dec 04 11:48:10 crc kubenswrapper[4979]: I1204 11:48:10.865435 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" podUID="6b7bb127-9196-4baa-b581-5ec01aa6da33" containerName="controller-manager" containerID="cri-o://126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0" gracePeriod=30 Dec 04 11:48:10 crc kubenswrapper[4979]: I1204 11:48:10.968347 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2"] Dec 04 11:48:10 crc kubenswrapper[4979]: I1204 11:48:10.968696 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" podUID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" containerName="route-controller-manager" containerID="cri-o://957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf" gracePeriod=30 Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.198420 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.257473 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273147 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-config\") pod \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273488 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b7bb127-9196-4baa-b581-5ec01aa6da33-serving-cert\") pod \"6b7bb127-9196-4baa-b581-5ec01aa6da33\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273515 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-config\") pod \"6b7bb127-9196-4baa-b581-5ec01aa6da33\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273543 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr6kb\" (UniqueName: \"kubernetes.io/projected/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-kube-api-access-kr6kb\") pod \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273576 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-proxy-ca-bundles\") pod \"6b7bb127-9196-4baa-b581-5ec01aa6da33\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273591 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-client-ca\") pod \"6b7bb127-9196-4baa-b581-5ec01aa6da33\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273617 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv4n9\" (UniqueName: \"kubernetes.io/projected/6b7bb127-9196-4baa-b581-5ec01aa6da33-kube-api-access-tv4n9\") pod \"6b7bb127-9196-4baa-b581-5ec01aa6da33\" (UID: \"6b7bb127-9196-4baa-b581-5ec01aa6da33\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273639 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-client-ca\") pod \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.273657 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-serving-cert\") pod \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\" (UID: \"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc\") " Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.274551 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-client-ca" (OuterVolumeSpecName: "client-ca") pod "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" (UID: "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.274850 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-config" (OuterVolumeSpecName: "config") pod "6b7bb127-9196-4baa-b581-5ec01aa6da33" (UID: "6b7bb127-9196-4baa-b581-5ec01aa6da33"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.274993 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-client-ca" (OuterVolumeSpecName: "client-ca") pod "6b7bb127-9196-4baa-b581-5ec01aa6da33" (UID: "6b7bb127-9196-4baa-b581-5ec01aa6da33"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.275329 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6b7bb127-9196-4baa-b581-5ec01aa6da33" (UID: "6b7bb127-9196-4baa-b581-5ec01aa6da33"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.278953 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-config" (OuterVolumeSpecName: "config") pod "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" (UID: "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.279852 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b7bb127-9196-4baa-b581-5ec01aa6da33-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6b7bb127-9196-4baa-b581-5ec01aa6da33" (UID: "6b7bb127-9196-4baa-b581-5ec01aa6da33"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.279931 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" (UID: "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.279962 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-kube-api-access-kr6kb" (OuterVolumeSpecName: "kube-api-access-kr6kb") pod "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" (UID: "17eb3c3d-c64c-416e-aa5f-c6839db5d1fc"). InnerVolumeSpecName "kube-api-access-kr6kb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.280941 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7bb127-9196-4baa-b581-5ec01aa6da33-kube-api-access-tv4n9" (OuterVolumeSpecName: "kube-api-access-tv4n9") pod "6b7bb127-9196-4baa-b581-5ec01aa6da33" (UID: "6b7bb127-9196-4baa-b581-5ec01aa6da33"). InnerVolumeSpecName "kube-api-access-tv4n9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374496 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv4n9\" (UniqueName: \"kubernetes.io/projected/6b7bb127-9196-4baa-b581-5ec01aa6da33-kube-api-access-tv4n9\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374546 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374559 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374572 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374586 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374597 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6b7bb127-9196-4baa-b581-5ec01aa6da33-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374609 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr6kb\" (UniqueName: \"kubernetes.io/projected/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc-kube-api-access-kr6kb\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374620 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.374631 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6b7bb127-9196-4baa-b581-5ec01aa6da33-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.811206 4979 generic.go:334] "Generic (PLEG): container finished" podID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" containerID="957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf" exitCode=0 Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.811546 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" event={"ID":"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc","Type":"ContainerDied","Data":"957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf"} Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.811601 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.811624 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" event={"ID":"17eb3c3d-c64c-416e-aa5f-c6839db5d1fc","Type":"ContainerDied","Data":"276980da103f03402f3dd3d08dfeecb7b182abceb049c8ec29f7cb1c1782f329"} Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.811691 4979 scope.go:117] "RemoveContainer" containerID="957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.814922 4979 generic.go:334] "Generic (PLEG): container finished" podID="6b7bb127-9196-4baa-b581-5ec01aa6da33" containerID="126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0" exitCode=0 Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.815009 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" event={"ID":"6b7bb127-9196-4baa-b581-5ec01aa6da33","Type":"ContainerDied","Data":"126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0"} Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.815056 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.815087 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" event={"ID":"6b7bb127-9196-4baa-b581-5ec01aa6da33","Type":"ContainerDied","Data":"c71b7b3f12fe147f15593e4c52da47276281e51c392589c45517edc06fc9e94e"} Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.838614 4979 scope.go:117] "RemoveContainer" containerID="957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf" Dec 04 11:48:11 crc kubenswrapper[4979]: E1204 11:48:11.839241 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf\": container with ID starting with 957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf not found: ID does not exist" containerID="957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.839285 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf"} err="failed to get container status \"957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf\": rpc error: code = NotFound desc = could not find container \"957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf\": container with ID starting with 957eb107e2b96ea16548f1b121a0f23e39b4b6ffba3fa6ec0d3b705542c771bf not found: ID does not exist" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.839348 4979 scope.go:117] "RemoveContainer" containerID="126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.864017 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhnlh"] Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.864493 4979 scope.go:117] "RemoveContainer" containerID="126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0" Dec 04 11:48:11 crc kubenswrapper[4979]: E1204 11:48:11.865011 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0\": container with ID starting with 126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0 not found: ID does not exist" containerID="126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.865065 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0"} err="failed to get container status \"126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0\": rpc error: code = NotFound desc = could not find container \"126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0\": container with ID starting with 126ed2e8f19625726bd992ca1fe563f6eed56e16f225947cb929c3b806d646b0 not found: ID does not exist" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.868714 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-bhnlh"] Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.871695 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2"] Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.874511 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2"] Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.942069 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74996bb75-nrngp"] Dec 04 11:48:11 crc kubenswrapper[4979]: E1204 11:48:11.942322 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" containerName="route-controller-manager" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.942339 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" containerName="route-controller-manager" Dec 04 11:48:11 crc kubenswrapper[4979]: E1204 11:48:11.942355 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7bb127-9196-4baa-b581-5ec01aa6da33" containerName="controller-manager" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.942364 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7bb127-9196-4baa-b581-5ec01aa6da33" containerName="controller-manager" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.942475 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" containerName="route-controller-manager" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.942493 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7bb127-9196-4baa-b581-5ec01aa6da33" containerName="controller-manager" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.942922 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.949964 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.950327 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.950776 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.951062 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.951064 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.954624 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.956482 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74996bb75-nrngp"] Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.958134 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.982266 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9hwc\" (UniqueName: \"kubernetes.io/projected/dee3b74f-5083-4cfa-9aba-bcffe41654b8-kube-api-access-g9hwc\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.982397 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-client-ca\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.982553 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-config\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.982604 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dee3b74f-5083-4cfa-9aba-bcffe41654b8-serving-cert\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:11 crc kubenswrapper[4979]: I1204 11:48:11.982669 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-proxy-ca-bundles\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.083897 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-client-ca\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.083982 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-config\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.084017 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dee3b74f-5083-4cfa-9aba-bcffe41654b8-serving-cert\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.084078 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-proxy-ca-bundles\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.084115 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9hwc\" (UniqueName: \"kubernetes.io/projected/dee3b74f-5083-4cfa-9aba-bcffe41654b8-kube-api-access-g9hwc\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.085015 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-client-ca\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.085641 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-proxy-ca-bundles\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.086938 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-config\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.090025 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dee3b74f-5083-4cfa-9aba-bcffe41654b8-serving-cert\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.103530 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9hwc\" (UniqueName: \"kubernetes.io/projected/dee3b74f-5083-4cfa-9aba-bcffe41654b8-kube-api-access-g9hwc\") pod \"controller-manager-74996bb75-nrngp\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.105212 4979 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-bhnlh container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.105315 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-bhnlh" podUID="6b7bb127-9196-4baa-b581-5ec01aa6da33" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.165591 4979 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-mdvl2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.165674 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-mdvl2" podUID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.204934 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17eb3c3d-c64c-416e-aa5f-c6839db5d1fc" path="/var/lib/kubelet/pods/17eb3c3d-c64c-416e-aa5f-c6839db5d1fc/volumes" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.205522 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7bb127-9196-4baa-b581-5ec01aa6da33" path="/var/lib/kubelet/pods/6b7bb127-9196-4baa-b581-5ec01aa6da33/volumes" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.275427 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.462370 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74996bb75-nrngp"] Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.822342 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" event={"ID":"dee3b74f-5083-4cfa-9aba-bcffe41654b8","Type":"ContainerStarted","Data":"85e3d587c03bdd3eeb593ba30b14bbe90b9b67cc4bcb4105bced31b692caf0a8"} Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.822397 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" event={"ID":"dee3b74f-5083-4cfa-9aba-bcffe41654b8","Type":"ContainerStarted","Data":"2f83113b5d59438daddc569e891e0e6814ef5c92c51512cd634e1f4509877652"} Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.822516 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.827487 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.840723 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" podStartSLOduration=2.840702099 podStartE2EDuration="2.840702099s" podCreationTimestamp="2025-12-04 11:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:48:12.838259608 +0000 UTC m=+317.112555412" watchObservedRunningTime="2025-12-04 11:48:12.840702099 +0000 UTC m=+317.114997903" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.943309 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh"] Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.944383 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.947730 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.947933 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.947958 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.948055 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.948122 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.949008 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.953606 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh"] Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.995929 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-config\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.996012 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b46bc22-4618-400d-a5e5-b343fffbf7eb-serving-cert\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.996066 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6549\" (UniqueName: \"kubernetes.io/projected/8b46bc22-4618-400d-a5e5-b343fffbf7eb-kube-api-access-q6549\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:12 crc kubenswrapper[4979]: I1204 11:48:12.996110 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-client-ca\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.097332 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-config\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.097479 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b46bc22-4618-400d-a5e5-b343fffbf7eb-serving-cert\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.097554 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6549\" (UniqueName: \"kubernetes.io/projected/8b46bc22-4618-400d-a5e5-b343fffbf7eb-kube-api-access-q6549\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.097617 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-client-ca\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.098597 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-client-ca\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.099019 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-config\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.105902 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b46bc22-4618-400d-a5e5-b343fffbf7eb-serving-cert\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.114707 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6549\" (UniqueName: \"kubernetes.io/projected/8b46bc22-4618-400d-a5e5-b343fffbf7eb-kube-api-access-q6549\") pod \"route-controller-manager-bc9f44c49-scwwh\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.304460 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.506067 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh"] Dec 04 11:48:13 crc kubenswrapper[4979]: W1204 11:48:13.518265 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b46bc22_4618_400d_a5e5_b343fffbf7eb.slice/crio-1e681edfdd78771e6ee6ff8004456789847df8651e49e9063835d26ef04bb906 WatchSource:0}: Error finding container 1e681edfdd78771e6ee6ff8004456789847df8651e49e9063835d26ef04bb906: Status 404 returned error can't find the container with id 1e681edfdd78771e6ee6ff8004456789847df8651e49e9063835d26ef04bb906 Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.832256 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" event={"ID":"8b46bc22-4618-400d-a5e5-b343fffbf7eb","Type":"ContainerStarted","Data":"4b83320a44deb8b03a07fd0b3924f58be9d65255d31d6cf7060d47ca9b417490"} Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.832290 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" event={"ID":"8b46bc22-4618-400d-a5e5-b343fffbf7eb","Type":"ContainerStarted","Data":"1e681edfdd78771e6ee6ff8004456789847df8651e49e9063835d26ef04bb906"} Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.832549 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.862035 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" podStartSLOduration=3.86195528 podStartE2EDuration="3.86195528s" podCreationTimestamp="2025-12-04 11:48:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:48:13.848582124 +0000 UTC m=+318.122877938" watchObservedRunningTime="2025-12-04 11:48:13.86195528 +0000 UTC m=+318.136251124" Dec 04 11:48:13 crc kubenswrapper[4979]: I1204 11:48:13.910536 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:28 crc kubenswrapper[4979]: I1204 11:48:28.040962 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:48:28 crc kubenswrapper[4979]: I1204 11:48:28.041509 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.761932 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74996bb75-nrngp"] Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.762433 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" podUID="dee3b74f-5083-4cfa-9aba-bcffe41654b8" containerName="controller-manager" containerID="cri-o://85e3d587c03bdd3eeb593ba30b14bbe90b9b67cc4bcb4105bced31b692caf0a8" gracePeriod=30 Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.775366 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh"] Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.775570 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" podUID="8b46bc22-4618-400d-a5e5-b343fffbf7eb" containerName="route-controller-manager" containerID="cri-o://4b83320a44deb8b03a07fd0b3924f58be9d65255d31d6cf7060d47ca9b417490" gracePeriod=30 Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.926099 4979 generic.go:334] "Generic (PLEG): container finished" podID="dee3b74f-5083-4cfa-9aba-bcffe41654b8" containerID="85e3d587c03bdd3eeb593ba30b14bbe90b9b67cc4bcb4105bced31b692caf0a8" exitCode=0 Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.926280 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" event={"ID":"dee3b74f-5083-4cfa-9aba-bcffe41654b8","Type":"ContainerDied","Data":"85e3d587c03bdd3eeb593ba30b14bbe90b9b67cc4bcb4105bced31b692caf0a8"} Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.928312 4979 generic.go:334] "Generic (PLEG): container finished" podID="8b46bc22-4618-400d-a5e5-b343fffbf7eb" containerID="4b83320a44deb8b03a07fd0b3924f58be9d65255d31d6cf7060d47ca9b417490" exitCode=0 Dec 04 11:48:30 crc kubenswrapper[4979]: I1204 11:48:30.928349 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" event={"ID":"8b46bc22-4618-400d-a5e5-b343fffbf7eb","Type":"ContainerDied","Data":"4b83320a44deb8b03a07fd0b3924f58be9d65255d31d6cf7060d47ca9b417490"} Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.247894 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.311988 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.415764 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dee3b74f-5083-4cfa-9aba-bcffe41654b8-serving-cert\") pod \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.415814 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-client-ca\") pod \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.415853 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-config\") pod \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.415877 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b46bc22-4618-400d-a5e5-b343fffbf7eb-serving-cert\") pod \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.415927 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q6549\" (UniqueName: \"kubernetes.io/projected/8b46bc22-4618-400d-a5e5-b343fffbf7eb-kube-api-access-q6549\") pod \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\" (UID: \"8b46bc22-4618-400d-a5e5-b343fffbf7eb\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.415976 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-client-ca\") pod \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.416105 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-proxy-ca-bundles\") pod \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.416173 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-config\") pod \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.416269 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9hwc\" (UniqueName: \"kubernetes.io/projected/dee3b74f-5083-4cfa-9aba-bcffe41654b8-kube-api-access-g9hwc\") pod \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\" (UID: \"dee3b74f-5083-4cfa-9aba-bcffe41654b8\") " Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.416694 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-client-ca" (OuterVolumeSpecName: "client-ca") pod "dee3b74f-5083-4cfa-9aba-bcffe41654b8" (UID: "dee3b74f-5083-4cfa-9aba-bcffe41654b8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.416821 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "dee3b74f-5083-4cfa-9aba-bcffe41654b8" (UID: "dee3b74f-5083-4cfa-9aba-bcffe41654b8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.416920 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-config" (OuterVolumeSpecName: "config") pod "dee3b74f-5083-4cfa-9aba-bcffe41654b8" (UID: "dee3b74f-5083-4cfa-9aba-bcffe41654b8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.417004 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-config" (OuterVolumeSpecName: "config") pod "8b46bc22-4618-400d-a5e5-b343fffbf7eb" (UID: "8b46bc22-4618-400d-a5e5-b343fffbf7eb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.417133 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-client-ca" (OuterVolumeSpecName: "client-ca") pod "8b46bc22-4618-400d-a5e5-b343fffbf7eb" (UID: "8b46bc22-4618-400d-a5e5-b343fffbf7eb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.417207 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.417230 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.417243 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.417257 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/dee3b74f-5083-4cfa-9aba-bcffe41654b8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.421370 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b46bc22-4618-400d-a5e5-b343fffbf7eb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8b46bc22-4618-400d-a5e5-b343fffbf7eb" (UID: "8b46bc22-4618-400d-a5e5-b343fffbf7eb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.421599 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dee3b74f-5083-4cfa-9aba-bcffe41654b8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dee3b74f-5083-4cfa-9aba-bcffe41654b8" (UID: "dee3b74f-5083-4cfa-9aba-bcffe41654b8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.421483 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b46bc22-4618-400d-a5e5-b343fffbf7eb-kube-api-access-q6549" (OuterVolumeSpecName: "kube-api-access-q6549") pod "8b46bc22-4618-400d-a5e5-b343fffbf7eb" (UID: "8b46bc22-4618-400d-a5e5-b343fffbf7eb"). InnerVolumeSpecName "kube-api-access-q6549". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.421512 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dee3b74f-5083-4cfa-9aba-bcffe41654b8-kube-api-access-g9hwc" (OuterVolumeSpecName: "kube-api-access-g9hwc") pod "dee3b74f-5083-4cfa-9aba-bcffe41654b8" (UID: "dee3b74f-5083-4cfa-9aba-bcffe41654b8"). InnerVolumeSpecName "kube-api-access-g9hwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.518197 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9hwc\" (UniqueName: \"kubernetes.io/projected/dee3b74f-5083-4cfa-9aba-bcffe41654b8-kube-api-access-g9hwc\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.518232 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dee3b74f-5083-4cfa-9aba-bcffe41654b8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.518242 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8b46bc22-4618-400d-a5e5-b343fffbf7eb-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.518251 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b46bc22-4618-400d-a5e5-b343fffbf7eb-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.518260 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q6549\" (UniqueName: \"kubernetes.io/projected/8b46bc22-4618-400d-a5e5-b343fffbf7eb-kube-api-access-q6549\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.936958 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.936952 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74996bb75-nrngp" event={"ID":"dee3b74f-5083-4cfa-9aba-bcffe41654b8","Type":"ContainerDied","Data":"2f83113b5d59438daddc569e891e0e6814ef5c92c51512cd634e1f4509877652"} Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.937147 4979 scope.go:117] "RemoveContainer" containerID="85e3d587c03bdd3eeb593ba30b14bbe90b9b67cc4bcb4105bced31b692caf0a8" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.939798 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" event={"ID":"8b46bc22-4618-400d-a5e5-b343fffbf7eb","Type":"ContainerDied","Data":"1e681edfdd78771e6ee6ff8004456789847df8651e49e9063835d26ef04bb906"} Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.939948 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.964628 4979 scope.go:117] "RemoveContainer" containerID="4b83320a44deb8b03a07fd0b3924f58be9d65255d31d6cf7060d47ca9b417490" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.971536 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x"] Dec 04 11:48:31 crc kubenswrapper[4979]: E1204 11:48:31.972629 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dee3b74f-5083-4cfa-9aba-bcffe41654b8" containerName="controller-manager" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.972646 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dee3b74f-5083-4cfa-9aba-bcffe41654b8" containerName="controller-manager" Dec 04 11:48:31 crc kubenswrapper[4979]: E1204 11:48:31.972673 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b46bc22-4618-400d-a5e5-b343fffbf7eb" containerName="route-controller-manager" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.972680 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b46bc22-4618-400d-a5e5-b343fffbf7eb" containerName="route-controller-manager" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.972992 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dee3b74f-5083-4cfa-9aba-bcffe41654b8" containerName="controller-manager" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.973012 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b46bc22-4618-400d-a5e5-b343fffbf7eb" containerName="route-controller-manager" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.976966 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.984888 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.985024 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6"] Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.985232 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.985529 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.985926 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.987695 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x"] Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.985658 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.986116 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.986182 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.990864 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6"] Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.992591 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.992688 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.992920 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.993034 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.993346 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.994602 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 04 11:48:31 crc kubenswrapper[4979]: I1204 11:48:31.998431 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024237 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-config\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024283 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-client-ca\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024317 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tw4jv\" (UniqueName: \"kubernetes.io/projected/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-kube-api-access-tw4jv\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024336 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktgcw\" (UniqueName: \"kubernetes.io/projected/b4db635e-ef3f-450c-a840-b79d107eaf2f-kube-api-access-ktgcw\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024353 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-client-ca\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024382 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4db635e-ef3f-450c-a840-b79d107eaf2f-serving-cert\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024410 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-config\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024424 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-serving-cert\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.024446 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-proxy-ca-bundles\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.030055 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74996bb75-nrngp"] Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.037125 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74996bb75-nrngp"] Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.040810 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh"] Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.046802 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bc9f44c49-scwwh"] Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126109 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4db635e-ef3f-450c-a840-b79d107eaf2f-serving-cert\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126178 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-config\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126206 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-serving-cert\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126239 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-proxy-ca-bundles\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126269 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-config\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126320 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-client-ca\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126348 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tw4jv\" (UniqueName: \"kubernetes.io/projected/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-kube-api-access-tw4jv\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126380 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktgcw\" (UniqueName: \"kubernetes.io/projected/b4db635e-ef3f-450c-a840-b79d107eaf2f-kube-api-access-ktgcw\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.126413 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-client-ca\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.127986 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-client-ca\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.128257 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-config\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.128688 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-client-ca\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.129155 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-proxy-ca-bundles\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.129381 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-config\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.132035 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4db635e-ef3f-450c-a840-b79d107eaf2f-serving-cert\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.132093 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-serving-cert\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.148445 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktgcw\" (UniqueName: \"kubernetes.io/projected/b4db635e-ef3f-450c-a840-b79d107eaf2f-kube-api-access-ktgcw\") pod \"controller-manager-74bcd77bc4-4nx6x\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.148903 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tw4jv\" (UniqueName: \"kubernetes.io/projected/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-kube-api-access-tw4jv\") pod \"route-controller-manager-659875ffb8-r84g6\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.204857 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b46bc22-4618-400d-a5e5-b343fffbf7eb" path="/var/lib/kubelet/pods/8b46bc22-4618-400d-a5e5-b343fffbf7eb/volumes" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.205378 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dee3b74f-5083-4cfa-9aba-bcffe41654b8" path="/var/lib/kubelet/pods/dee3b74f-5083-4cfa-9aba-bcffe41654b8/volumes" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.318428 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.326679 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.522969 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6"] Dec 04 11:48:32 crc kubenswrapper[4979]: W1204 11:48:32.529253 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e3ee1fd_46eb_4d7d_86d6_13191126cf61.slice/crio-fb6c5da7b17b432945992040a85fc1463ff895d6214273ad04596e509345f084 WatchSource:0}: Error finding container fb6c5da7b17b432945992040a85fc1463ff895d6214273ad04596e509345f084: Status 404 returned error can't find the container with id fb6c5da7b17b432945992040a85fc1463ff895d6214273ad04596e509345f084 Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.703306 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x"] Dec 04 11:48:32 crc kubenswrapper[4979]: W1204 11:48:32.711021 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4db635e_ef3f_450c_a840_b79d107eaf2f.slice/crio-63da71bc4588a57878ce31cc36ed3a30cd89e8e494ec2f19869fdc936a17364f WatchSource:0}: Error finding container 63da71bc4588a57878ce31cc36ed3a30cd89e8e494ec2f19869fdc936a17364f: Status 404 returned error can't find the container with id 63da71bc4588a57878ce31cc36ed3a30cd89e8e494ec2f19869fdc936a17364f Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.946686 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" event={"ID":"0e3ee1fd-46eb-4d7d-86d6-13191126cf61","Type":"ContainerStarted","Data":"cbff29ca03d5eed85416da582123d58a09ce12c48170d70c9d0b0ecd52fe5d5a"} Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.947067 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.947084 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" event={"ID":"0e3ee1fd-46eb-4d7d-86d6-13191126cf61","Type":"ContainerStarted","Data":"fb6c5da7b17b432945992040a85fc1463ff895d6214273ad04596e509345f084"} Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.950317 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" event={"ID":"b4db635e-ef3f-450c-a840-b79d107eaf2f","Type":"ContainerStarted","Data":"a77db135e8916f8277b9aef23f8a1ee5820d3e726756ffa9b528acb85997e487"} Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.950377 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" event={"ID":"b4db635e-ef3f-450c-a840-b79d107eaf2f","Type":"ContainerStarted","Data":"63da71bc4588a57878ce31cc36ed3a30cd89e8e494ec2f19869fdc936a17364f"} Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.950500 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.960577 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" podStartSLOduration=2.960558911 podStartE2EDuration="2.960558911s" podCreationTimestamp="2025-12-04 11:48:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:48:32.960042156 +0000 UTC m=+337.234337960" watchObservedRunningTime="2025-12-04 11:48:32.960558911 +0000 UTC m=+337.234854715" Dec 04 11:48:32 crc kubenswrapper[4979]: I1204 11:48:32.961790 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:48:33 crc kubenswrapper[4979]: I1204 11:48:33.001437 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" podStartSLOduration=3.001418163 podStartE2EDuration="3.001418163s" podCreationTimestamp="2025-12-04 11:48:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:48:32.980374494 +0000 UTC m=+337.254670298" watchObservedRunningTime="2025-12-04 11:48:33.001418163 +0000 UTC m=+337.275713977" Dec 04 11:48:33 crc kubenswrapper[4979]: I1204 11:48:33.165740 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:50 crc kubenswrapper[4979]: I1204 11:48:50.875038 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6"] Dec 04 11:48:50 crc kubenswrapper[4979]: I1204 11:48:50.875779 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" podUID="0e3ee1fd-46eb-4d7d-86d6-13191126cf61" containerName="route-controller-manager" containerID="cri-o://cbff29ca03d5eed85416da582123d58a09ce12c48170d70c9d0b0ecd52fe5d5a" gracePeriod=30 Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.062412 4979 generic.go:334] "Generic (PLEG): container finished" podID="0e3ee1fd-46eb-4d7d-86d6-13191126cf61" containerID="cbff29ca03d5eed85416da582123d58a09ce12c48170d70c9d0b0ecd52fe5d5a" exitCode=0 Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.062455 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" event={"ID":"0e3ee1fd-46eb-4d7d-86d6-13191126cf61","Type":"ContainerDied","Data":"cbff29ca03d5eed85416da582123d58a09ce12c48170d70c9d0b0ecd52fe5d5a"} Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.332237 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.471248 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tw4jv\" (UniqueName: \"kubernetes.io/projected/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-kube-api-access-tw4jv\") pod \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.471340 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-config\") pod \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.471377 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-client-ca\") pod \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.471395 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-serving-cert\") pod \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\" (UID: \"0e3ee1fd-46eb-4d7d-86d6-13191126cf61\") " Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.472969 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-client-ca" (OuterVolumeSpecName: "client-ca") pod "0e3ee1fd-46eb-4d7d-86d6-13191126cf61" (UID: "0e3ee1fd-46eb-4d7d-86d6-13191126cf61"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.473084 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-config" (OuterVolumeSpecName: "config") pod "0e3ee1fd-46eb-4d7d-86d6-13191126cf61" (UID: "0e3ee1fd-46eb-4d7d-86d6-13191126cf61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.476882 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-kube-api-access-tw4jv" (OuterVolumeSpecName: "kube-api-access-tw4jv") pod "0e3ee1fd-46eb-4d7d-86d6-13191126cf61" (UID: "0e3ee1fd-46eb-4d7d-86d6-13191126cf61"). InnerVolumeSpecName "kube-api-access-tw4jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.477195 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0e3ee1fd-46eb-4d7d-86d6-13191126cf61" (UID: "0e3ee1fd-46eb-4d7d-86d6-13191126cf61"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.573073 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tw4jv\" (UniqueName: \"kubernetes.io/projected/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-kube-api-access-tw4jv\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.573122 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.573136 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.573148 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0e3ee1fd-46eb-4d7d-86d6-13191126cf61-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.974146 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm"] Dec 04 11:48:51 crc kubenswrapper[4979]: E1204 11:48:51.974413 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e3ee1fd-46eb-4d7d-86d6-13191126cf61" containerName="route-controller-manager" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.974431 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e3ee1fd-46eb-4d7d-86d6-13191126cf61" containerName="route-controller-manager" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.974525 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e3ee1fd-46eb-4d7d-86d6-13191126cf61" containerName="route-controller-manager" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.974862 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:51 crc kubenswrapper[4979]: I1204 11:48:51.990510 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm"] Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.068910 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" event={"ID":"0e3ee1fd-46eb-4d7d-86d6-13191126cf61","Type":"ContainerDied","Data":"fb6c5da7b17b432945992040a85fc1463ff895d6214273ad04596e509345f084"} Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.068962 4979 scope.go:117] "RemoveContainer" containerID="cbff29ca03d5eed85416da582123d58a09ce12c48170d70c9d0b0ecd52fe5d5a" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.068961 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.079582 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc22e781-c65b-4241-95c5-ff396991fd8a-serving-cert\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.079640 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dc22e781-c65b-4241-95c5-ff396991fd8a-client-ca\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.079824 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd7pg\" (UniqueName: \"kubernetes.io/projected/dc22e781-c65b-4241-95c5-ff396991fd8a-kube-api-access-sd7pg\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.079894 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc22e781-c65b-4241-95c5-ff396991fd8a-config\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.098391 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6"] Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.107375 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-659875ffb8-r84g6"] Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.181183 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd7pg\" (UniqueName: \"kubernetes.io/projected/dc22e781-c65b-4241-95c5-ff396991fd8a-kube-api-access-sd7pg\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.181235 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc22e781-c65b-4241-95c5-ff396991fd8a-config\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.181289 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc22e781-c65b-4241-95c5-ff396991fd8a-serving-cert\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.181336 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dc22e781-c65b-4241-95c5-ff396991fd8a-client-ca\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.182480 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dc22e781-c65b-4241-95c5-ff396991fd8a-config\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.182611 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/dc22e781-c65b-4241-95c5-ff396991fd8a-client-ca\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.184644 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dc22e781-c65b-4241-95c5-ff396991fd8a-serving-cert\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.204510 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e3ee1fd-46eb-4d7d-86d6-13191126cf61" path="/var/lib/kubelet/pods/0e3ee1fd-46eb-4d7d-86d6-13191126cf61/volumes" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.212112 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd7pg\" (UniqueName: \"kubernetes.io/projected/dc22e781-c65b-4241-95c5-ff396991fd8a-kube-api-access-sd7pg\") pod \"route-controller-manager-7f85f5db49-vlhjm\" (UID: \"dc22e781-c65b-4241-95c5-ff396991fd8a\") " pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.312186 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:52 crc kubenswrapper[4979]: I1204 11:48:52.700703 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm"] Dec 04 11:48:53 crc kubenswrapper[4979]: I1204 11:48:53.075253 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" event={"ID":"dc22e781-c65b-4241-95c5-ff396991fd8a","Type":"ContainerStarted","Data":"16732702ebb833a77a07efa51738386527c518236f65227fad8f32c15697be30"} Dec 04 11:48:53 crc kubenswrapper[4979]: I1204 11:48:53.075739 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:53 crc kubenswrapper[4979]: I1204 11:48:53.075769 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" event={"ID":"dc22e781-c65b-4241-95c5-ff396991fd8a","Type":"ContainerStarted","Data":"a9e57f1f33b8a62578ffa86a2590b89784395ee38c48acca67aafb94e4bf7a97"} Dec 04 11:48:53 crc kubenswrapper[4979]: I1204 11:48:53.095408 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" podStartSLOduration=3.095390155 podStartE2EDuration="3.095390155s" podCreationTimestamp="2025-12-04 11:48:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:48:53.093651135 +0000 UTC m=+357.367946939" watchObservedRunningTime="2025-12-04 11:48:53.095390155 +0000 UTC m=+357.369685959" Dec 04 11:48:53 crc kubenswrapper[4979]: I1204 11:48:53.329851 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7f85f5db49-vlhjm" Dec 04 11:48:58 crc kubenswrapper[4979]: I1204 11:48:58.040794 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:48:58 crc kubenswrapper[4979]: I1204 11:48:58.041511 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.607125 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cm5bl"] Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.608278 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.618582 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cm5bl"] Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.769931 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.769997 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7b6s\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-kube-api-access-g7b6s\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.770015 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5966baf6-6444-49b2-8f31-1d3baa3337e7-registry-certificates\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.770043 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-bound-sa-token\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.770070 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5966baf6-6444-49b2-8f31-1d3baa3337e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.770091 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5966baf6-6444-49b2-8f31-1d3baa3337e7-trusted-ca\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.770108 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-registry-tls\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.770128 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5966baf6-6444-49b2-8f31-1d3baa3337e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.796371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871169 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7b6s\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-kube-api-access-g7b6s\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871217 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5966baf6-6444-49b2-8f31-1d3baa3337e7-registry-certificates\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871243 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-bound-sa-token\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871279 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5966baf6-6444-49b2-8f31-1d3baa3337e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871328 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5966baf6-6444-49b2-8f31-1d3baa3337e7-trusted-ca\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871423 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-registry-tls\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871451 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5966baf6-6444-49b2-8f31-1d3baa3337e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.871959 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5966baf6-6444-49b2-8f31-1d3baa3337e7-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.872890 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5966baf6-6444-49b2-8f31-1d3baa3337e7-trusted-ca\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.872981 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5966baf6-6444-49b2-8f31-1d3baa3337e7-registry-certificates\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.883105 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5966baf6-6444-49b2-8f31-1d3baa3337e7-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.883218 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-registry-tls\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.887352 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7b6s\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-kube-api-access-g7b6s\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.887653 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5966baf6-6444-49b2-8f31-1d3baa3337e7-bound-sa-token\") pod \"image-registry-66df7c8f76-cm5bl\" (UID: \"5966baf6-6444-49b2-8f31-1d3baa3337e7\") " pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:04 crc kubenswrapper[4979]: I1204 11:49:04.923918 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:05 crc kubenswrapper[4979]: I1204 11:49:05.340161 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cm5bl"] Dec 04 11:49:06 crc kubenswrapper[4979]: I1204 11:49:06.138188 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" event={"ID":"5966baf6-6444-49b2-8f31-1d3baa3337e7","Type":"ContainerStarted","Data":"b657ba6e913271bb21eb5273d723b6ebe4b5e78d757d3fe57d1203530f6d5369"} Dec 04 11:49:06 crc kubenswrapper[4979]: I1204 11:49:06.138844 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:06 crc kubenswrapper[4979]: I1204 11:49:06.138899 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" event={"ID":"5966baf6-6444-49b2-8f31-1d3baa3337e7","Type":"ContainerStarted","Data":"ba67b2051b324e410bbd052205d4bd63136db12ee0197a4304ab7580f75915d4"} Dec 04 11:49:06 crc kubenswrapper[4979]: I1204 11:49:06.157707 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" podStartSLOduration=2.157682603 podStartE2EDuration="2.157682603s" podCreationTimestamp="2025-12-04 11:49:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:49:06.154191772 +0000 UTC m=+370.428487576" watchObservedRunningTime="2025-12-04 11:49:06.157682603 +0000 UTC m=+370.431978407" Dec 04 11:49:10 crc kubenswrapper[4979]: I1204 11:49:10.866597 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x"] Dec 04 11:49:10 crc kubenswrapper[4979]: I1204 11:49:10.867560 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" podUID="b4db635e-ef3f-450c-a840-b79d107eaf2f" containerName="controller-manager" containerID="cri-o://a77db135e8916f8277b9aef23f8a1ee5820d3e726756ffa9b528acb85997e487" gracePeriod=30 Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.171018 4979 generic.go:334] "Generic (PLEG): container finished" podID="b4db635e-ef3f-450c-a840-b79d107eaf2f" containerID="a77db135e8916f8277b9aef23f8a1ee5820d3e726756ffa9b528acb85997e487" exitCode=0 Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.171067 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" event={"ID":"b4db635e-ef3f-450c-a840-b79d107eaf2f","Type":"ContainerDied","Data":"a77db135e8916f8277b9aef23f8a1ee5820d3e726756ffa9b528acb85997e487"} Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.235064 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.361728 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-proxy-ca-bundles\") pod \"b4db635e-ef3f-450c-a840-b79d107eaf2f\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.361807 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktgcw\" (UniqueName: \"kubernetes.io/projected/b4db635e-ef3f-450c-a840-b79d107eaf2f-kube-api-access-ktgcw\") pod \"b4db635e-ef3f-450c-a840-b79d107eaf2f\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.361852 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4db635e-ef3f-450c-a840-b79d107eaf2f-serving-cert\") pod \"b4db635e-ef3f-450c-a840-b79d107eaf2f\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.361870 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-config\") pod \"b4db635e-ef3f-450c-a840-b79d107eaf2f\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.361905 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-client-ca\") pod \"b4db635e-ef3f-450c-a840-b79d107eaf2f\" (UID: \"b4db635e-ef3f-450c-a840-b79d107eaf2f\") " Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.362626 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-client-ca" (OuterVolumeSpecName: "client-ca") pod "b4db635e-ef3f-450c-a840-b79d107eaf2f" (UID: "b4db635e-ef3f-450c-a840-b79d107eaf2f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.363130 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b4db635e-ef3f-450c-a840-b79d107eaf2f" (UID: "b4db635e-ef3f-450c-a840-b79d107eaf2f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.363466 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-config" (OuterVolumeSpecName: "config") pod "b4db635e-ef3f-450c-a840-b79d107eaf2f" (UID: "b4db635e-ef3f-450c-a840-b79d107eaf2f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.367607 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4db635e-ef3f-450c-a840-b79d107eaf2f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b4db635e-ef3f-450c-a840-b79d107eaf2f" (UID: "b4db635e-ef3f-450c-a840-b79d107eaf2f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.367787 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4db635e-ef3f-450c-a840-b79d107eaf2f-kube-api-access-ktgcw" (OuterVolumeSpecName: "kube-api-access-ktgcw") pod "b4db635e-ef3f-450c-a840-b79d107eaf2f" (UID: "b4db635e-ef3f-450c-a840-b79d107eaf2f"). InnerVolumeSpecName "kube-api-access-ktgcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.463212 4979 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.463253 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktgcw\" (UniqueName: \"kubernetes.io/projected/b4db635e-ef3f-450c-a840-b79d107eaf2f-kube-api-access-ktgcw\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.463268 4979 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b4db635e-ef3f-450c-a840-b79d107eaf2f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.463278 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.463286 4979 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b4db635e-ef3f-450c-a840-b79d107eaf2f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.983543 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d7854dd95-prqbc"] Dec 04 11:49:11 crc kubenswrapper[4979]: E1204 11:49:11.983769 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4db635e-ef3f-450c-a840-b79d107eaf2f" containerName="controller-manager" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.983784 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4db635e-ef3f-450c-a840-b79d107eaf2f" containerName="controller-manager" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.983897 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4db635e-ef3f-450c-a840-b79d107eaf2f" containerName="controller-manager" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.984351 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:11 crc kubenswrapper[4979]: I1204 11:49:11.998927 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d7854dd95-prqbc"] Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.071499 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-proxy-ca-bundles\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.071585 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-client-ca\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.071721 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms72l\" (UniqueName: \"kubernetes.io/projected/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-kube-api-access-ms72l\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.071831 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-config\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.071894 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-serving-cert\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.173106 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-proxy-ca-bundles\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.173203 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-client-ca\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.173274 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms72l\" (UniqueName: \"kubernetes.io/projected/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-kube-api-access-ms72l\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.173393 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-config\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.173442 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-serving-cert\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.174264 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-client-ca\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.174999 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-proxy-ca-bundles\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.176484 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-config\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.180000 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-serving-cert\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.184237 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" event={"ID":"b4db635e-ef3f-450c-a840-b79d107eaf2f","Type":"ContainerDied","Data":"63da71bc4588a57878ce31cc36ed3a30cd89e8e494ec2f19869fdc936a17364f"} Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.184357 4979 scope.go:117] "RemoveContainer" containerID="a77db135e8916f8277b9aef23f8a1ee5820d3e726756ffa9b528acb85997e487" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.184483 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.191037 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms72l\" (UniqueName: \"kubernetes.io/projected/1aa586a1-ea1b-4cd1-abf2-abad5a10b089-kube-api-access-ms72l\") pod \"controller-manager-5d7854dd95-prqbc\" (UID: \"1aa586a1-ea1b-4cd1-abf2-abad5a10b089\") " pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.263261 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x"] Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.270049 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74bcd77bc4-4nx6x"] Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.373387 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:12 crc kubenswrapper[4979]: I1204 11:49:12.759707 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d7854dd95-prqbc"] Dec 04 11:49:13 crc kubenswrapper[4979]: I1204 11:49:13.191536 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" event={"ID":"1aa586a1-ea1b-4cd1-abf2-abad5a10b089","Type":"ContainerStarted","Data":"f54ee0ca84df67f2ddb854476be77aeb799ae0158fee5071d93e15b65d3aab51"} Dec 04 11:49:13 crc kubenswrapper[4979]: I1204 11:49:13.191584 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" event={"ID":"1aa586a1-ea1b-4cd1-abf2-abad5a10b089","Type":"ContainerStarted","Data":"bd47379214374dd8dec36d516ed18488918ca81f974a1c631ad4d174a14ea144"} Dec 04 11:49:13 crc kubenswrapper[4979]: I1204 11:49:13.191820 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:13 crc kubenswrapper[4979]: I1204 11:49:13.201938 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" Dec 04 11:49:13 crc kubenswrapper[4979]: I1204 11:49:13.211372 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d7854dd95-prqbc" podStartSLOduration=3.211351304 podStartE2EDuration="3.211351304s" podCreationTimestamp="2025-12-04 11:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:49:13.208822361 +0000 UTC m=+377.483118175" watchObservedRunningTime="2025-12-04 11:49:13.211351304 +0000 UTC m=+377.485647108" Dec 04 11:49:14 crc kubenswrapper[4979]: I1204 11:49:14.211460 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4db635e-ef3f-450c-a840-b79d107eaf2f" path="/var/lib/kubelet/pods/b4db635e-ef3f-450c-a840-b79d107eaf2f/volumes" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.476591 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5k8fs"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.477917 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5k8fs" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="registry-server" containerID="cri-o://1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a" gracePeriod=30 Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.496293 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kwwbv"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.497103 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kwwbv" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="registry-server" containerID="cri-o://0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419" gracePeriod=30 Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.498914 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csp56"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.499121 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" containerID="cri-o://d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961" gracePeriod=30 Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.503704 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6zl"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.504086 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vz6zl" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="registry-server" containerID="cri-o://74e33fafe09603ba2cf538b1a6ecbbc353545446a71e08f9475176e2693a27ae" gracePeriod=30 Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.518233 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gqrbf"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.518527 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gqrbf" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="registry-server" containerID="cri-o://6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695" gracePeriod=30 Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.531163 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mvpph"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.532035 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.540369 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mvpph"] Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.718072 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54723683-07ae-4574-9396-43fc24d977af-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.718481 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsvcc\" (UniqueName: \"kubernetes.io/projected/54723683-07ae-4574-9396-43fc24d977af-kube-api-access-xsvcc\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.718504 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54723683-07ae-4574-9396-43fc24d977af-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.819528 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54723683-07ae-4574-9396-43fc24d977af-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.819643 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsvcc\" (UniqueName: \"kubernetes.io/projected/54723683-07ae-4574-9396-43fc24d977af-kube-api-access-xsvcc\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.819674 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54723683-07ae-4574-9396-43fc24d977af-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.821221 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/54723683-07ae-4574-9396-43fc24d977af-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.828371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/54723683-07ae-4574-9396-43fc24d977af-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.839697 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsvcc\" (UniqueName: \"kubernetes.io/projected/54723683-07ae-4574-9396-43fc24d977af-kube-api-access-xsvcc\") pod \"marketplace-operator-79b997595-mvpph\" (UID: \"54723683-07ae-4574-9396-43fc24d977af\") " pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:22 crc kubenswrapper[4979]: I1204 11:49:22.851465 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.005568 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.125911 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-catalog-content\") pod \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.126374 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzmtf\" (UniqueName: \"kubernetes.io/projected/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-kube-api-access-xzmtf\") pod \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.126409 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-utilities\") pod \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\" (UID: \"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.127466 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-utilities" (OuterVolumeSpecName: "utilities") pod "c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" (UID: "c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.130040 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-kube-api-access-xzmtf" (OuterVolumeSpecName: "kube-api-access-xzmtf") pod "c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" (UID: "c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea"). InnerVolumeSpecName "kube-api-access-xzmtf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.183830 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" (UID: "c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.228094 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzmtf\" (UniqueName: \"kubernetes.io/projected/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-kube-api-access-xzmtf\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.228129 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.228141 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.242633 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.256468 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.272256 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.279185 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerID="6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695" exitCode=0 Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.279219 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gqrbf" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.279331 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqrbf" event={"ID":"bc617bb8-ffc1-48b8-9342-136d54548e2e","Type":"ContainerDied","Data":"6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.279365 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gqrbf" event={"ID":"bc617bb8-ffc1-48b8-9342-136d54548e2e","Type":"ContainerDied","Data":"dfc6b31300abde26e873e0a57ef5a962c7dce359e5832b9457603ad5f3189b15"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.279388 4979 scope.go:117] "RemoveContainer" containerID="6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.295430 4979 generic.go:334] "Generic (PLEG): container finished" podID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerID="74e33fafe09603ba2cf538b1a6ecbbc353545446a71e08f9475176e2693a27ae" exitCode=0 Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.295610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6zl" event={"ID":"190bcf1a-d610-48d5-a8b5-d94f2369b8d3","Type":"ContainerDied","Data":"74e33fafe09603ba2cf538b1a6ecbbc353545446a71e08f9475176e2693a27ae"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.299280 4979 generic.go:334] "Generic (PLEG): container finished" podID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerID="1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a" exitCode=0 Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.299362 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5k8fs" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.299391 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k8fs" event={"ID":"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea","Type":"ContainerDied","Data":"1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.299423 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5k8fs" event={"ID":"c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea","Type":"ContainerDied","Data":"ab51031a284b7fd248de011320b6e3f8882a43e51d5c136357e2ec17c4f08f97"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.303921 4979 generic.go:334] "Generic (PLEG): container finished" podID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerID="0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419" exitCode=0 Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.304006 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwwbv" event={"ID":"96fc77cc-ab61-462b-9e6a-0877bb125da6","Type":"ContainerDied","Data":"0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.304054 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kwwbv" event={"ID":"96fc77cc-ab61-462b-9e6a-0877bb125da6","Type":"ContainerDied","Data":"3ffc07c5303641ccfa8b0d4c71c28d1ad12050bc99566477caf003c43d01d396"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.304104 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kwwbv" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.311675 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.313256 4979 scope.go:117] "RemoveContainer" containerID="63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.317272 4979 generic.go:334] "Generic (PLEG): container finished" podID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerID="d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961" exitCode=0 Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.317327 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" event={"ID":"16cbfdaf-1b4c-463e-9767-6879dc885056","Type":"ContainerDied","Data":"d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.317350 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" event={"ID":"16cbfdaf-1b4c-463e-9767-6879dc885056","Type":"ContainerDied","Data":"29db5cbe3770f70c2bcb572272301f2146d5c1d641b0fda1af33d476fe6b3faa"} Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.317405 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-csp56" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.350594 4979 scope.go:117] "RemoveContainer" containerID="e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.358507 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5k8fs"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.362719 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5k8fs"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.366230 4979 scope.go:117] "RemoveContainer" containerID="6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.366724 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695\": container with ID starting with 6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695 not found: ID does not exist" containerID="6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.366798 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695"} err="failed to get container status \"6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695\": rpc error: code = NotFound desc = could not find container \"6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695\": container with ID starting with 6541dc100c26c69278e251d182c63fd166af7d0644d8bc66c50bc93e56d17695 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.366829 4979 scope.go:117] "RemoveContainer" containerID="63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.367200 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979\": container with ID starting with 63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979 not found: ID does not exist" containerID="63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.367232 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979"} err="failed to get container status \"63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979\": rpc error: code = NotFound desc = could not find container \"63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979\": container with ID starting with 63464ee382bc4617cf58cafd6c96c90df2c80dd701ccab3bf6e72a05a404a979 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.367262 4979 scope.go:117] "RemoveContainer" containerID="e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.367582 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f\": container with ID starting with e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f not found: ID does not exist" containerID="e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.367619 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f"} err="failed to get container status \"e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f\": rpc error: code = NotFound desc = could not find container \"e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f\": container with ID starting with e670dc7be8edaf3bff470c71a36c36e3442abd025e0daa9311c9557b6041f62f not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.367659 4979 scope.go:117] "RemoveContainer" containerID="1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.380498 4979 scope.go:117] "RemoveContainer" containerID="878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.394284 4979 scope.go:117] "RemoveContainer" containerID="b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.417939 4979 scope.go:117] "RemoveContainer" containerID="1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.418648 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a\": container with ID starting with 1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a not found: ID does not exist" containerID="1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.418687 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a"} err="failed to get container status \"1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a\": rpc error: code = NotFound desc = could not find container \"1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a\": container with ID starting with 1c03cf32dbcc12401b786bde25aed9e0c0d25b60bf20018c14f1e9b53b52c66a not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.418718 4979 scope.go:117] "RemoveContainer" containerID="878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.419441 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4\": container with ID starting with 878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4 not found: ID does not exist" containerID="878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.419503 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4"} err="failed to get container status \"878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4\": rpc error: code = NotFound desc = could not find container \"878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4\": container with ID starting with 878b92eed4869e97864d35a935deda818266fb18f255ec82c4ff5517a6c5daf4 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.419540 4979 scope.go:117] "RemoveContainer" containerID="b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.420003 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66\": container with ID starting with b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66 not found: ID does not exist" containerID="b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.420062 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66"} err="failed to get container status \"b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66\": rpc error: code = NotFound desc = could not find container \"b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66\": container with ID starting with b0286b47ab4333a4ec3d3837f24d08569246c11454035bccb9c825e5e8662d66 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.420082 4979 scope.go:117] "RemoveContainer" containerID="0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431075 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-catalog-content\") pod \"96fc77cc-ab61-462b-9e6a-0877bb125da6\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431105 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6ptw\" (UniqueName: \"kubernetes.io/projected/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-kube-api-access-x6ptw\") pod \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431145 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-utilities\") pod \"bc617bb8-ffc1-48b8-9342-136d54548e2e\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431182 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-operator-metrics\") pod \"16cbfdaf-1b4c-463e-9767-6879dc885056\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431201 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-trusted-ca\") pod \"16cbfdaf-1b4c-463e-9767-6879dc885056\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431233 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-catalog-content\") pod \"bc617bb8-ffc1-48b8-9342-136d54548e2e\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431262 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-utilities\") pod \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431292 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbc9f\" (UniqueName: \"kubernetes.io/projected/16cbfdaf-1b4c-463e-9767-6879dc885056-kube-api-access-fbc9f\") pod \"16cbfdaf-1b4c-463e-9767-6879dc885056\" (UID: \"16cbfdaf-1b4c-463e-9767-6879dc885056\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431327 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx5c8\" (UniqueName: \"kubernetes.io/projected/bc617bb8-ffc1-48b8-9342-136d54548e2e-kube-api-access-mx5c8\") pod \"bc617bb8-ffc1-48b8-9342-136d54548e2e\" (UID: \"bc617bb8-ffc1-48b8-9342-136d54548e2e\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431349 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hl5vc\" (UniqueName: \"kubernetes.io/projected/96fc77cc-ab61-462b-9e6a-0877bb125da6-kube-api-access-hl5vc\") pod \"96fc77cc-ab61-462b-9e6a-0877bb125da6\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431370 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-utilities\") pod \"96fc77cc-ab61-462b-9e6a-0877bb125da6\" (UID: \"96fc77cc-ab61-462b-9e6a-0877bb125da6\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.431395 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-catalog-content\") pod \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\" (UID: \"190bcf1a-d610-48d5-a8b5-d94f2369b8d3\") " Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.434736 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-utilities" (OuterVolumeSpecName: "utilities") pod "190bcf1a-d610-48d5-a8b5-d94f2369b8d3" (UID: "190bcf1a-d610-48d5-a8b5-d94f2369b8d3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.434845 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-utilities" (OuterVolumeSpecName: "utilities") pod "96fc77cc-ab61-462b-9e6a-0877bb125da6" (UID: "96fc77cc-ab61-462b-9e6a-0877bb125da6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.435442 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "16cbfdaf-1b4c-463e-9767-6879dc885056" (UID: "16cbfdaf-1b4c-463e-9767-6879dc885056"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.438217 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "16cbfdaf-1b4c-463e-9767-6879dc885056" (UID: "16cbfdaf-1b4c-463e-9767-6879dc885056"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.440613 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-utilities" (OuterVolumeSpecName: "utilities") pod "bc617bb8-ffc1-48b8-9342-136d54548e2e" (UID: "bc617bb8-ffc1-48b8-9342-136d54548e2e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.440934 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.442336 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.442366 4979 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.442393 4979 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16cbfdaf-1b4c-463e-9767-6879dc885056-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.442407 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.447094 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-kube-api-access-x6ptw" (OuterVolumeSpecName: "kube-api-access-x6ptw") pod "190bcf1a-d610-48d5-a8b5-d94f2369b8d3" (UID: "190bcf1a-d610-48d5-a8b5-d94f2369b8d3"). InnerVolumeSpecName "kube-api-access-x6ptw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.449968 4979 scope.go:117] "RemoveContainer" containerID="f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.456443 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16cbfdaf-1b4c-463e-9767-6879dc885056-kube-api-access-fbc9f" (OuterVolumeSpecName: "kube-api-access-fbc9f") pod "16cbfdaf-1b4c-463e-9767-6879dc885056" (UID: "16cbfdaf-1b4c-463e-9767-6879dc885056"). InnerVolumeSpecName "kube-api-access-fbc9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.456927 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96fc77cc-ab61-462b-9e6a-0877bb125da6-kube-api-access-hl5vc" (OuterVolumeSpecName: "kube-api-access-hl5vc") pod "96fc77cc-ab61-462b-9e6a-0877bb125da6" (UID: "96fc77cc-ab61-462b-9e6a-0877bb125da6"). InnerVolumeSpecName "kube-api-access-hl5vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.457708 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc617bb8-ffc1-48b8-9342-136d54548e2e-kube-api-access-mx5c8" (OuterVolumeSpecName: "kube-api-access-mx5c8") pod "bc617bb8-ffc1-48b8-9342-136d54548e2e" (UID: "bc617bb8-ffc1-48b8-9342-136d54548e2e"). InnerVolumeSpecName "kube-api-access-mx5c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.472631 4979 scope.go:117] "RemoveContainer" containerID="013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.482088 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "190bcf1a-d610-48d5-a8b5-d94f2369b8d3" (UID: "190bcf1a-d610-48d5-a8b5-d94f2369b8d3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.511399 4979 scope.go:117] "RemoveContainer" containerID="0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.512272 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419\": container with ID starting with 0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419 not found: ID does not exist" containerID="0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.512429 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419"} err="failed to get container status \"0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419\": rpc error: code = NotFound desc = could not find container \"0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419\": container with ID starting with 0e8213b7d749e14e4755db0cc59257e3c767b0c1c9ad08bef038348c3977f419 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.512602 4979 scope.go:117] "RemoveContainer" containerID="f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.513006 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe\": container with ID starting with f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe not found: ID does not exist" containerID="f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.513127 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe"} err="failed to get container status \"f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe\": rpc error: code = NotFound desc = could not find container \"f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe\": container with ID starting with f1949bf04ce77e44e0c7a162a2525d7398bf6009aa3e386368fecedd13d165fe not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.513234 4979 scope.go:117] "RemoveContainer" containerID="013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.513580 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50\": container with ID starting with 013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50 not found: ID does not exist" containerID="013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.513686 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50"} err="failed to get container status \"013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50\": rpc error: code = NotFound desc = could not find container \"013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50\": container with ID starting with 013a0b954f6d4c0d2980ebf8f3ae7c70feea0e2a5f23fa703e51da8c4b153e50 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.513804 4979 scope.go:117] "RemoveContainer" containerID="d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.514471 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-mvpph"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.533615 4979 scope.go:117] "RemoveContainer" containerID="6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.534628 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96fc77cc-ab61-462b-9e6a-0877bb125da6" (UID: "96fc77cc-ab61-462b-9e6a-0877bb125da6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.543803 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbc9f\" (UniqueName: \"kubernetes.io/projected/16cbfdaf-1b4c-463e-9767-6879dc885056-kube-api-access-fbc9f\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.543899 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx5c8\" (UniqueName: \"kubernetes.io/projected/bc617bb8-ffc1-48b8-9342-136d54548e2e-kube-api-access-mx5c8\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.543912 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl5vc\" (UniqueName: \"kubernetes.io/projected/96fc77cc-ab61-462b-9e6a-0877bb125da6-kube-api-access-hl5vc\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.543926 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.543938 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96fc77cc-ab61-462b-9e6a-0877bb125da6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.543948 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6ptw\" (UniqueName: \"kubernetes.io/projected/190bcf1a-d610-48d5-a8b5-d94f2369b8d3-kube-api-access-x6ptw\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.548086 4979 scope.go:117] "RemoveContainer" containerID="d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.548913 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961\": container with ID starting with d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961 not found: ID does not exist" containerID="d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.549248 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961"} err="failed to get container status \"d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961\": rpc error: code = NotFound desc = could not find container \"d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961\": container with ID starting with d2fe2ab49f34e15aad71192e15b07628927aa971867eead9ddf1ec3d47214961 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.549291 4979 scope.go:117] "RemoveContainer" containerID="6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804" Dec 04 11:49:23 crc kubenswrapper[4979]: E1204 11:49:23.549719 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804\": container with ID starting with 6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804 not found: ID does not exist" containerID="6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.549747 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804"} err="failed to get container status \"6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804\": rpc error: code = NotFound desc = could not find container \"6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804\": container with ID starting with 6f9a350add74f3847fb41a38566807f29076f07533d417c3152c97f86f5a0804 not found: ID does not exist" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.587044 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc617bb8-ffc1-48b8-9342-136d54548e2e" (UID: "bc617bb8-ffc1-48b8-9342-136d54548e2e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.645715 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc617bb8-ffc1-48b8-9342-136d54548e2e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.688018 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kwwbv"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.692921 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kwwbv"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.697149 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csp56"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.699895 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-csp56"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.905150 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gqrbf"] Dec 04 11:49:23 crc kubenswrapper[4979]: I1204 11:49:23.908317 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gqrbf"] Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.205173 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" path="/var/lib/kubelet/pods/16cbfdaf-1b4c-463e-9767-6879dc885056/volumes" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.205665 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" path="/var/lib/kubelet/pods/96fc77cc-ab61-462b-9e6a-0877bb125da6/volumes" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.206273 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" path="/var/lib/kubelet/pods/bc617bb8-ffc1-48b8-9342-136d54548e2e/volumes" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.206856 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" path="/var/lib/kubelet/pods/c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea/volumes" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.324992 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vz6zl" event={"ID":"190bcf1a-d610-48d5-a8b5-d94f2369b8d3","Type":"ContainerDied","Data":"8040c58669939a1fc6176da42b154811deb95d7c0216c880f326c4af3aca73d9"} Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.325352 4979 scope.go:117] "RemoveContainer" containerID="74e33fafe09603ba2cf538b1a6ecbbc353545446a71e08f9475176e2693a27ae" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.325022 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vz6zl" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.329379 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" event={"ID":"54723683-07ae-4574-9396-43fc24d977af","Type":"ContainerStarted","Data":"a96438a473c20c8e8308b4c55e4c083fd3af7e97938e460bc569b76756f708c1"} Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.329415 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" event={"ID":"54723683-07ae-4574-9396-43fc24d977af","Type":"ContainerStarted","Data":"2992d220ac3921f63f9b2f475c6237319efd376e2a08bc2f0c40171a588fa854"} Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.330150 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.333371 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.343897 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6zl"] Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.347828 4979 scope.go:117] "RemoveContainer" containerID="d6b9a4d4b768af8b19aa043d5837ffc8f1bd8282e78d3ad96d50db1d1d465b00" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.353500 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vz6zl"] Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.360873 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-mvpph" podStartSLOduration=2.360856097 podStartE2EDuration="2.360856097s" podCreationTimestamp="2025-12-04 11:49:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:49:24.358841569 +0000 UTC m=+388.633137373" watchObservedRunningTime="2025-12-04 11:49:24.360856097 +0000 UTC m=+388.635151901" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.373060 4979 scope.go:117] "RemoveContainer" containerID="5679b3fa1b5f7dad8e903a8b1c102fa7bcfab64e4ff690151b54b87fe1e74801" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.899419 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bdrvn"] Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.899955 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.899974 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.899991 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900060 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900073 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900104 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900116 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900123 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900134 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900143 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900152 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900159 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900190 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900198 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900205 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900211 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900218 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900226 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="extract-utilities" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900237 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900243 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900277 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900284 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900292 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900328 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.900339 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900346 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="extract-content" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900509 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900524 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="96fc77cc-ab61-462b-9e6a-0877bb125da6" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900534 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc617bb8-ffc1-48b8-9342-136d54548e2e" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900570 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.900582 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f95b37-c4e3-46e2-a027-4cd8fd6fc9ea" containerName="registry-server" Dec 04 11:49:24 crc kubenswrapper[4979]: E1204 11:49:24.901191 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.901204 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.901376 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="16cbfdaf-1b4c-463e-9767-6879dc885056" containerName="marketplace-operator" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.902607 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.905312 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.907416 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdrvn"] Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.928578 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-cm5bl" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.965671 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6359145-7b66-40db-833e-4bb609833266-catalog-content\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.966082 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6359145-7b66-40db-833e-4bb609833266-utilities\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.966266 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82b5c\" (UniqueName: \"kubernetes.io/projected/a6359145-7b66-40db-833e-4bb609833266-kube-api-access-82b5c\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:24 crc kubenswrapper[4979]: I1204 11:49:24.982342 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-stbh4"] Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.067661 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6359145-7b66-40db-833e-4bb609833266-catalog-content\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.067817 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6359145-7b66-40db-833e-4bb609833266-utilities\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.067886 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82b5c\" (UniqueName: \"kubernetes.io/projected/a6359145-7b66-40db-833e-4bb609833266-kube-api-access-82b5c\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.068561 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a6359145-7b66-40db-833e-4bb609833266-catalog-content\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.068739 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a6359145-7b66-40db-833e-4bb609833266-utilities\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.085456 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82b5c\" (UniqueName: \"kubernetes.io/projected/a6359145-7b66-40db-833e-4bb609833266-kube-api-access-82b5c\") pod \"redhat-operators-bdrvn\" (UID: \"a6359145-7b66-40db-833e-4bb609833266\") " pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.223410 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:25 crc kubenswrapper[4979]: I1204 11:49:25.608990 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bdrvn"] Dec 04 11:49:25 crc kubenswrapper[4979]: W1204 11:49:25.622973 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6359145_7b66_40db_833e_4bb609833266.slice/crio-2c6c600fe0e399e21c0896557bd8dcd37fb8273d0ac82869833db5956c263337 WatchSource:0}: Error finding container 2c6c600fe0e399e21c0896557bd8dcd37fb8273d0ac82869833db5956c263337: Status 404 returned error can't find the container with id 2c6c600fe0e399e21c0896557bd8dcd37fb8273d0ac82869833db5956c263337 Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.206617 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="190bcf1a-d610-48d5-a8b5-d94f2369b8d3" path="/var/lib/kubelet/pods/190bcf1a-d610-48d5-a8b5-d94f2369b8d3/volumes" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.357223 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6359145-7b66-40db-833e-4bb609833266" containerID="0add9492c27008cc43d412ecb59e92d6545667d4c17f08713884fa786c5e067c" exitCode=0 Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.357336 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdrvn" event={"ID":"a6359145-7b66-40db-833e-4bb609833266","Type":"ContainerDied","Data":"0add9492c27008cc43d412ecb59e92d6545667d4c17f08713884fa786c5e067c"} Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.357428 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdrvn" event={"ID":"a6359145-7b66-40db-833e-4bb609833266","Type":"ContainerStarted","Data":"2c6c600fe0e399e21c0896557bd8dcd37fb8273d0ac82869833db5956c263337"} Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.694221 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r8q2n"] Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.695885 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.699426 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.705069 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r8q2n"] Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.890706 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-utilities\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.890754 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r45xn\" (UniqueName: \"kubernetes.io/projected/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-kube-api-access-r45xn\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.890797 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-catalog-content\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.992452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-utilities\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.992507 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r45xn\" (UniqueName: \"kubernetes.io/projected/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-kube-api-access-r45xn\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.992554 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-catalog-content\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.993032 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-utilities\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:26 crc kubenswrapper[4979]: I1204 11:49:26.993052 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-catalog-content\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.014943 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r45xn\" (UniqueName: \"kubernetes.io/projected/028b7d1f-52bf-4bc2-8905-2c1d5035b3fc-kube-api-access-r45xn\") pod \"certified-operators-r8q2n\" (UID: \"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc\") " pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.296696 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d7n8l"] Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.300419 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.304271 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.306071 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7n8l"] Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.311147 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.375499 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdrvn" event={"ID":"a6359145-7b66-40db-833e-4bb609833266","Type":"ContainerStarted","Data":"0bb96926ce0c7aae88a15ef33821f17815b35ed1221236fbf0a425825309439d"} Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.498052 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-utilities\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.498155 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ffgl\" (UniqueName: \"kubernetes.io/projected/28921e2c-1159-4b88-bfbf-0214efef7819-kube-api-access-6ffgl\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.498182 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-catalog-content\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.599748 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-utilities\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.599813 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6ffgl\" (UniqueName: \"kubernetes.io/projected/28921e2c-1159-4b88-bfbf-0214efef7819-kube-api-access-6ffgl\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.599841 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-catalog-content\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.600367 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-utilities\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.600394 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-catalog-content\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.621639 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6ffgl\" (UniqueName: \"kubernetes.io/projected/28921e2c-1159-4b88-bfbf-0214efef7819-kube-api-access-6ffgl\") pod \"community-operators-d7n8l\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.627456 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:27 crc kubenswrapper[4979]: I1204 11:49:27.785058 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r8q2n"] Dec 04 11:49:27 crc kubenswrapper[4979]: W1204 11:49:27.821067 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod028b7d1f_52bf_4bc2_8905_2c1d5035b3fc.slice/crio-d0488806573dad63a1a15a77fa351695722ad2611969cabb53cdcff8c132ec0e WatchSource:0}: Error finding container d0488806573dad63a1a15a77fa351695722ad2611969cabb53cdcff8c132ec0e: Status 404 returned error can't find the container with id d0488806573dad63a1a15a77fa351695722ad2611969cabb53cdcff8c132ec0e Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.040954 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.041241 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.041288 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.041936 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e61fe1ef8585786eb9e6a5ae2992ed20ce3863f21b72946cbe3aff92d1a5bf00"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.042001 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://e61fe1ef8585786eb9e6a5ae2992ed20ce3863f21b72946cbe3aff92d1a5bf00" gracePeriod=600 Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.070113 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d7n8l"] Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.386269 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6359145-7b66-40db-833e-4bb609833266" containerID="0bb96926ce0c7aae88a15ef33821f17815b35ed1221236fbf0a425825309439d" exitCode=0 Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.386350 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdrvn" event={"ID":"a6359145-7b66-40db-833e-4bb609833266","Type":"ContainerDied","Data":"0bb96926ce0c7aae88a15ef33821f17815b35ed1221236fbf0a425825309439d"} Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.389001 4979 generic.go:334] "Generic (PLEG): container finished" podID="028b7d1f-52bf-4bc2-8905-2c1d5035b3fc" containerID="2b706045388937da85b22798f55593f619c2ffce6e18ec4c9025300826231168" exitCode=0 Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.389057 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8q2n" event={"ID":"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc","Type":"ContainerDied","Data":"2b706045388937da85b22798f55593f619c2ffce6e18ec4c9025300826231168"} Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.389107 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8q2n" event={"ID":"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc","Type":"ContainerStarted","Data":"d0488806573dad63a1a15a77fa351695722ad2611969cabb53cdcff8c132ec0e"} Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.392617 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="e61fe1ef8585786eb9e6a5ae2992ed20ce3863f21b72946cbe3aff92d1a5bf00" exitCode=0 Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.392692 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"e61fe1ef8585786eb9e6a5ae2992ed20ce3863f21b72946cbe3aff92d1a5bf00"} Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.392763 4979 scope.go:117] "RemoveContainer" containerID="c2dabaa99bda99d35e38ac06ec0e0e699d6eab59680ac3c1e9371bf8b706ee67" Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.399970 4979 generic.go:334] "Generic (PLEG): container finished" podID="28921e2c-1159-4b88-bfbf-0214efef7819" containerID="8c155757f1feb7bdbb352e5ee02882f2c51bcc8e47d840ae3653adaaa38a5108" exitCode=0 Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.400023 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerDied","Data":"8c155757f1feb7bdbb352e5ee02882f2c51bcc8e47d840ae3653adaaa38a5108"} Dec 04 11:49:28 crc kubenswrapper[4979]: I1204 11:49:28.400062 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerStarted","Data":"91329605ed1b4c95e32531b6ce834f968a09ac22c4a116797c82d4f8ee89e679"} Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.097431 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nfw4p"] Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.099521 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.101254 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.106688 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfw4p"] Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.175458 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c39667b-75fa-4174-896d-2424f435fde8-catalog-content\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.175671 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbz24\" (UniqueName: \"kubernetes.io/projected/3c39667b-75fa-4174-896d-2424f435fde8-kube-api-access-tbz24\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.175811 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c39667b-75fa-4174-896d-2424f435fde8-utilities\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.276583 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c39667b-75fa-4174-896d-2424f435fde8-catalog-content\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.276680 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbz24\" (UniqueName: \"kubernetes.io/projected/3c39667b-75fa-4174-896d-2424f435fde8-kube-api-access-tbz24\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.276742 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c39667b-75fa-4174-896d-2424f435fde8-utilities\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.278494 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c39667b-75fa-4174-896d-2424f435fde8-utilities\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.278623 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c39667b-75fa-4174-896d-2424f435fde8-catalog-content\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.326390 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbz24\" (UniqueName: \"kubernetes.io/projected/3c39667b-75fa-4174-896d-2424f435fde8-kube-api-access-tbz24\") pod \"redhat-marketplace-nfw4p\" (UID: \"3c39667b-75fa-4174-896d-2424f435fde8\") " pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.408951 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bdrvn" event={"ID":"a6359145-7b66-40db-833e-4bb609833266","Type":"ContainerStarted","Data":"8fb2499a2feb3c08d859b164f1802dfdb374968bf9d1c6b9fd2ff5d645568087"} Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.413606 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8q2n" event={"ID":"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc","Type":"ContainerStarted","Data":"b99f050b1994516a906c32d871ee28395e33e9742340fffbb83590ff4ea3350e"} Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.416048 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"ccf09364961bc412950c9a3b911d71a458a0f6d629773b900f305d69edd4ab19"} Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.417493 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerStarted","Data":"b67d4e615023233376bf96346977e618cfd637c0cf373f966265f9cd69781dfc"} Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.426761 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bdrvn" podStartSLOduration=2.967514367 podStartE2EDuration="5.426740938s" podCreationTimestamp="2025-12-04 11:49:24 +0000 UTC" firstStartedPulling="2025-12-04 11:49:26.359565335 +0000 UTC m=+390.633861139" lastFinishedPulling="2025-12-04 11:49:28.818791906 +0000 UTC m=+393.093087710" observedRunningTime="2025-12-04 11:49:29.423891569 +0000 UTC m=+393.698187403" watchObservedRunningTime="2025-12-04 11:49:29.426740938 +0000 UTC m=+393.701036742" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.452877 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:29 crc kubenswrapper[4979]: I1204 11:49:29.841473 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nfw4p"] Dec 04 11:49:29 crc kubenswrapper[4979]: W1204 11:49:29.860863 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c39667b_75fa_4174_896d_2424f435fde8.slice/crio-2f69a67be14880a893781ec7086f1e45cbc23026eba9203be5abfb94cb94fcaa WatchSource:0}: Error finding container 2f69a67be14880a893781ec7086f1e45cbc23026eba9203be5abfb94cb94fcaa: Status 404 returned error can't find the container with id 2f69a67be14880a893781ec7086f1e45cbc23026eba9203be5abfb94cb94fcaa Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.425253 4979 generic.go:334] "Generic (PLEG): container finished" podID="028b7d1f-52bf-4bc2-8905-2c1d5035b3fc" containerID="b99f050b1994516a906c32d871ee28395e33e9742340fffbb83590ff4ea3350e" exitCode=0 Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.425345 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8q2n" event={"ID":"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc","Type":"ContainerDied","Data":"b99f050b1994516a906c32d871ee28395e33e9742340fffbb83590ff4ea3350e"} Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.428509 4979 generic.go:334] "Generic (PLEG): container finished" podID="3c39667b-75fa-4174-896d-2424f435fde8" containerID="f7f3f4b89897ac39aa74491923b17e826ab652fb2cd25414f94e6a6b599ac2fb" exitCode=0 Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.428616 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfw4p" event={"ID":"3c39667b-75fa-4174-896d-2424f435fde8","Type":"ContainerDied","Data":"f7f3f4b89897ac39aa74491923b17e826ab652fb2cd25414f94e6a6b599ac2fb"} Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.428672 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfw4p" event={"ID":"3c39667b-75fa-4174-896d-2424f435fde8","Type":"ContainerStarted","Data":"2f69a67be14880a893781ec7086f1e45cbc23026eba9203be5abfb94cb94fcaa"} Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.432157 4979 generic.go:334] "Generic (PLEG): container finished" podID="28921e2c-1159-4b88-bfbf-0214efef7819" containerID="b67d4e615023233376bf96346977e618cfd637c0cf373f966265f9cd69781dfc" exitCode=0 Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.432213 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerDied","Data":"b67d4e615023233376bf96346977e618cfd637c0cf373f966265f9cd69781dfc"} Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.432241 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerStarted","Data":"22936e3142b9b6f3558f62c499f6164b3c0a8e615ffe30383fc5ef0368e42094"} Dec 04 11:49:30 crc kubenswrapper[4979]: I1204 11:49:30.466517 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d7n8l" podStartSLOduration=2.004410593 podStartE2EDuration="3.466499556s" podCreationTimestamp="2025-12-04 11:49:27 +0000 UTC" firstStartedPulling="2025-12-04 11:49:28.407286273 +0000 UTC m=+392.681582117" lastFinishedPulling="2025-12-04 11:49:29.869375276 +0000 UTC m=+394.143671080" observedRunningTime="2025-12-04 11:49:30.46377532 +0000 UTC m=+394.738071124" watchObservedRunningTime="2025-12-04 11:49:30.466499556 +0000 UTC m=+394.740795360" Dec 04 11:49:31 crc kubenswrapper[4979]: I1204 11:49:31.439781 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r8q2n" event={"ID":"028b7d1f-52bf-4bc2-8905-2c1d5035b3fc","Type":"ContainerStarted","Data":"4626e1682e1664b50d7e46f31d22b8a21f1f4d4e6fbbee517a04d674bc3dd758"} Dec 04 11:49:31 crc kubenswrapper[4979]: I1204 11:49:31.441534 4979 generic.go:334] "Generic (PLEG): container finished" podID="3c39667b-75fa-4174-896d-2424f435fde8" containerID="d7a0ce51ea8b2c3ac6da0f35f7991403238cba2c901b34a527ca2cd5316ea0b2" exitCode=0 Dec 04 11:49:31 crc kubenswrapper[4979]: I1204 11:49:31.441563 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfw4p" event={"ID":"3c39667b-75fa-4174-896d-2424f435fde8","Type":"ContainerDied","Data":"d7a0ce51ea8b2c3ac6da0f35f7991403238cba2c901b34a527ca2cd5316ea0b2"} Dec 04 11:49:31 crc kubenswrapper[4979]: I1204 11:49:31.464050 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r8q2n" podStartSLOduration=2.913464179 podStartE2EDuration="5.464027022s" podCreationTimestamp="2025-12-04 11:49:26 +0000 UTC" firstStartedPulling="2025-12-04 11:49:28.391022802 +0000 UTC m=+392.665318646" lastFinishedPulling="2025-12-04 11:49:30.941585685 +0000 UTC m=+395.215881489" observedRunningTime="2025-12-04 11:49:31.460651577 +0000 UTC m=+395.734947381" watchObservedRunningTime="2025-12-04 11:49:31.464027022 +0000 UTC m=+395.738322826" Dec 04 11:49:33 crc kubenswrapper[4979]: I1204 11:49:33.454003 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nfw4p" event={"ID":"3c39667b-75fa-4174-896d-2424f435fde8","Type":"ContainerStarted","Data":"0c9e1e9f371ad7aad058d410dc07518397579daee00028d237cfa7af4490f5a9"} Dec 04 11:49:33 crc kubenswrapper[4979]: I1204 11:49:33.470841 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nfw4p" podStartSLOduration=3.042115854 podStartE2EDuration="4.470821678s" podCreationTimestamp="2025-12-04 11:49:29 +0000 UTC" firstStartedPulling="2025-12-04 11:49:30.429521949 +0000 UTC m=+394.703817753" lastFinishedPulling="2025-12-04 11:49:31.858227763 +0000 UTC m=+396.132523577" observedRunningTime="2025-12-04 11:49:33.470729325 +0000 UTC m=+397.745025149" watchObservedRunningTime="2025-12-04 11:49:33.470821678 +0000 UTC m=+397.745117482" Dec 04 11:49:35 crc kubenswrapper[4979]: I1204 11:49:35.224627 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:35 crc kubenswrapper[4979]: I1204 11:49:35.225945 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:35 crc kubenswrapper[4979]: I1204 11:49:35.272040 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:35 crc kubenswrapper[4979]: I1204 11:49:35.502052 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bdrvn" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.312660 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.313190 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.359398 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.523986 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r8q2n" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.628567 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.628623 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:37 crc kubenswrapper[4979]: I1204 11:49:37.677512 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:38 crc kubenswrapper[4979]: I1204 11:49:38.521466 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 11:49:39 crc kubenswrapper[4979]: I1204 11:49:39.453726 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:39 crc kubenswrapper[4979]: I1204 11:49:39.453778 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:39 crc kubenswrapper[4979]: I1204 11:49:39.496457 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:39 crc kubenswrapper[4979]: I1204 11:49:39.539212 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nfw4p" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.029033 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" podUID="ac3bc99d-9384-4e8b-b266-92aa7a417150" containerName="registry" containerID="cri-o://413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb" gracePeriod=30 Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.434262 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.534557 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94s4r\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-kube-api-access-94s4r\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.534605 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-certificates\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.534636 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac3bc99d-9384-4e8b-b266-92aa7a417150-installation-pull-secrets\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.534662 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-bound-sa-token\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.534991 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.535019 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac3bc99d-9384-4e8b-b266-92aa7a417150-ca-trust-extracted\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.535077 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-tls\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.535115 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-trusted-ca\") pod \"ac3bc99d-9384-4e8b-b266-92aa7a417150\" (UID: \"ac3bc99d-9384-4e8b-b266-92aa7a417150\") " Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.535994 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.540221 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.540940 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.541162 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.542438 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-kube-api-access-94s4r" (OuterVolumeSpecName: "kube-api-access-94s4r") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "kube-api-access-94s4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.546318 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac3bc99d-9384-4e8b-b266-92aa7a417150-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.549100 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.554123 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac3bc99d-9384-4e8b-b266-92aa7a417150-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "ac3bc99d-9384-4e8b-b266-92aa7a417150" (UID: "ac3bc99d-9384-4e8b-b266-92aa7a417150"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.559955 4979 generic.go:334] "Generic (PLEG): container finished" podID="ac3bc99d-9384-4e8b-b266-92aa7a417150" containerID="413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb" exitCode=0 Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.560001 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" event={"ID":"ac3bc99d-9384-4e8b-b266-92aa7a417150","Type":"ContainerDied","Data":"413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb"} Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.560032 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" event={"ID":"ac3bc99d-9384-4e8b-b266-92aa7a417150","Type":"ContainerDied","Data":"1df2c3d3710577aa8c81e0476a1cec86af63a02121a9210c00a57f43c600968d"} Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.560056 4979 scope.go:117] "RemoveContainer" containerID="413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.560237 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-stbh4" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.595938 4979 scope.go:117] "RemoveContainer" containerID="413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb" Dec 04 11:49:50 crc kubenswrapper[4979]: E1204 11:49:50.596533 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb\": container with ID starting with 413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb not found: ID does not exist" containerID="413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.596574 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb"} err="failed to get container status \"413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb\": rpc error: code = NotFound desc = could not find container \"413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb\": container with ID starting with 413976aad34f794e0629b5c95f7cca76b4cc590508c9214d6afb229c153d90fb not found: ID does not exist" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.600063 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-stbh4"] Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.604628 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-stbh4"] Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636799 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94s4r\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-kube-api-access-94s4r\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636849 4979 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636874 4979 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ac3bc99d-9384-4e8b-b266-92aa7a417150-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636892 4979 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636909 4979 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ac3bc99d-9384-4e8b-b266-92aa7a417150-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636926 4979 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ac3bc99d-9384-4e8b-b266-92aa7a417150-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:50 crc kubenswrapper[4979]: I1204 11:49:50.636945 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ac3bc99d-9384-4e8b-b266-92aa7a417150-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:49:52 crc kubenswrapper[4979]: I1204 11:49:52.209458 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac3bc99d-9384-4e8b-b266-92aa7a417150" path="/var/lib/kubelet/pods/ac3bc99d-9384-4e8b-b266-92aa7a417150/volumes" Dec 04 11:51:28 crc kubenswrapper[4979]: I1204 11:51:28.041512 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:51:28 crc kubenswrapper[4979]: I1204 11:51:28.043975 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:51:58 crc kubenswrapper[4979]: I1204 11:51:58.041406 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:51:58 crc kubenswrapper[4979]: I1204 11:51:58.041936 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:52:28 crc kubenswrapper[4979]: I1204 11:52:28.041104 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:52:28 crc kubenswrapper[4979]: I1204 11:52:28.041757 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:52:28 crc kubenswrapper[4979]: I1204 11:52:28.041812 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:52:28 crc kubenswrapper[4979]: I1204 11:52:28.042387 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ccf09364961bc412950c9a3b911d71a458a0f6d629773b900f305d69edd4ab19"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 11:52:28 crc kubenswrapper[4979]: I1204 11:52:28.042449 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://ccf09364961bc412950c9a3b911d71a458a0f6d629773b900f305d69edd4ab19" gracePeriod=600 Dec 04 11:52:29 crc kubenswrapper[4979]: I1204 11:52:29.538464 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="ccf09364961bc412950c9a3b911d71a458a0f6d629773b900f305d69edd4ab19" exitCode=0 Dec 04 11:52:29 crc kubenswrapper[4979]: I1204 11:52:29.538664 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"ccf09364961bc412950c9a3b911d71a458a0f6d629773b900f305d69edd4ab19"} Dec 04 11:52:29 crc kubenswrapper[4979]: I1204 11:52:29.539052 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"aa49c902e467cec7eb7ad9fa7d11c15836d2d5553c53624c2c9cd8f02a5491c5"} Dec 04 11:52:29 crc kubenswrapper[4979]: I1204 11:52:29.539079 4979 scope.go:117] "RemoveContainer" containerID="e61fe1ef8585786eb9e6a5ae2992ed20ce3863f21b72946cbe3aff92d1a5bf00" Dec 04 11:54:58 crc kubenswrapper[4979]: I1204 11:54:58.040811 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:54:58 crc kubenswrapper[4979]: I1204 11:54:58.041401 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:55:25 crc kubenswrapper[4979]: I1204 11:55:25.484596 4979 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 04 11:55:28 crc kubenswrapper[4979]: I1204 11:55:28.041157 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:55:28 crc kubenswrapper[4979]: I1204 11:55:28.041604 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.040545 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.041115 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.041173 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.041920 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aa49c902e467cec7eb7ad9fa7d11c15836d2d5553c53624c2c9cd8f02a5491c5"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.041995 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://aa49c902e467cec7eb7ad9fa7d11c15836d2d5553c53624c2c9cd8f02a5491c5" gracePeriod=600 Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.786987 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="aa49c902e467cec7eb7ad9fa7d11c15836d2d5553c53624c2c9cd8f02a5491c5" exitCode=0 Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.787051 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"aa49c902e467cec7eb7ad9fa7d11c15836d2d5553c53624c2c9cd8f02a5491c5"} Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.787536 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d"} Dec 04 11:55:58 crc kubenswrapper[4979]: I1204 11:55:58.787557 4979 scope.go:117] "RemoveContainer" containerID="ccf09364961bc412950c9a3b911d71a458a0f6d629773b900f305d69edd4ab19" Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.857062 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x6k66"] Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.858113 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-controller" containerID="cri-o://a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.858612 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="sbdb" containerID="cri-o://516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.858698 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-node" containerID="cri-o://b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.858756 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-acl-logging" containerID="cri-o://d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.858726 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.858924 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="nbdb" containerID="cri-o://7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.859055 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="northd" containerID="cri-o://696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" gracePeriod=30 Dec 04 11:56:22 crc kubenswrapper[4979]: I1204 11:56:22.907216 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" containerID="cri-o://6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" gracePeriod=30 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.200634 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/3.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.204898 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovn-acl-logging/0.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.205597 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovn-controller/0.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.206072 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262348 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wd59t"] Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262528 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262539 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262550 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262555 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262563 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-acl-logging" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262569 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-acl-logging" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262577 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262583 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262588 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262594 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262602 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262607 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262615 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262623 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262638 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac3bc99d-9384-4e8b-b266-92aa7a417150" containerName="registry" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262649 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac3bc99d-9384-4e8b-b266-92aa7a417150" containerName="registry" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262662 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="northd" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262669 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="northd" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262679 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-node" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262686 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-node" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262694 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kubecfg-setup" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262701 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kubecfg-setup" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262710 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="nbdb" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262717 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="nbdb" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262729 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="sbdb" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262735 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="sbdb" Dec 04 11:56:23 crc kubenswrapper[4979]: E1204 11:56:23.262746 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262753 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262855 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-node" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262871 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="northd" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262882 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="sbdb" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262893 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262903 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262910 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="kube-rbac-proxy-ovn-metrics" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262917 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262924 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="nbdb" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262930 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac3bc99d-9384-4e8b-b266-92aa7a417150" containerName="registry" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262937 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.262946 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovn-acl-logging" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.263159 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.263176 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerName="ovnkube-controller" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.265120 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.345962 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-openvswitch\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346011 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-systemd\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346031 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-netd\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346051 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-etc-openvswitch\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346071 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-var-lib-openvswitch\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346098 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovn-node-metrics-cert\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346121 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-bin\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346145 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-systemd-units\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346164 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-env-overrides\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346182 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-ovn\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346196 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-config\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346215 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-kubelet\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346233 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-log-socket\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346250 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-script-lib\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346264 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-netns\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346277 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-slash\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346313 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-ovn-kubernetes\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346335 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h6d6\" (UniqueName: \"kubernetes.io/projected/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-kube-api-access-6h6d6\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346352 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-var-lib-cni-networks-ovn-kubernetes\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346368 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-node-log\") pod \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\" (UID: \"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26\") " Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346464 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346503 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-node-log" (OuterVolumeSpecName: "node-log") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346522 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-slash" (OuterVolumeSpecName: "host-slash") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346567 4979 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346577 4979 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-slash\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346585 4979 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-node-log\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346578 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346594 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-log-socket" (OuterVolumeSpecName: "log-socket") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346615 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346630 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346648 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346648 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346683 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346720 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.347344 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346755 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346779 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346856 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.346467 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.347096 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.351389 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.351797 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-kube-api-access-6h6d6" (OuterVolumeSpecName: "kube-api-access-6h6d6") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "kube-api-access-6h6d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.358416 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" (UID: "af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.448684 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-ovnkube-script-lib\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.448764 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-ovnkube-config\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.448827 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-run-netns\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.448875 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-run-ovn-kubernetes\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.448922 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-env-overrides\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449058 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-node-log\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449105 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02375d8f-59c3-418c-98e5-901917ef970f-ovn-node-metrics-cert\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449151 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ll4t\" (UniqueName: \"kubernetes.io/projected/02375d8f-59c3-418c-98e5-901917ef970f-kube-api-access-2ll4t\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449197 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449244 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449389 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-var-lib-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449458 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-ovn\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449493 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-systemd\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449523 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-slash\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449552 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-log-socket\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449584 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-cni-netd\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449624 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-kubelet\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449658 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-etc-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.449797 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-systemd-units\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-cni-bin\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450259 4979 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450369 4979 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450393 4979 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450446 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6h6d6\" (UniqueName: \"kubernetes.io/projected/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-kube-api-access-6h6d6\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450469 4979 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450489 4979 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450508 4979 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450524 4979 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450540 4979 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450557 4979 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450574 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450591 4979 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450608 4979 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450624 4979 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450640 4979 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450657 4979 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.450674 4979 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26-log-socket\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.550997 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ll4t\" (UniqueName: \"kubernetes.io/projected/02375d8f-59c3-418c-98e5-901917ef970f-kube-api-access-2ll4t\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551068 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551093 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551119 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-var-lib-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551143 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-ovn\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551165 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-systemd\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551183 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-slash\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551202 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-log-socket\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551222 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-cni-netd\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551248 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-kubelet\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551271 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-etc-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551273 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551310 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-ovn\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551294 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-systemd-units\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551339 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-systemd-units\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551343 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-log-socket\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551362 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-run-systemd\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551221 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-var-lib-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551275 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-cni-netd\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551368 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-slash\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551384 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-kubelet\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551249 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551448 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-etc-openvswitch\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551542 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-cni-bin\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551600 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-ovnkube-config\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551632 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-ovnkube-script-lib\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551663 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-run-netns\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551662 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-cni-bin\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551697 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-run-ovn-kubernetes\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551726 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-env-overrides\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551759 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02375d8f-59c3-418c-98e5-901917ef970f-ovn-node-metrics-cert\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551778 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-node-log\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551840 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-node-log\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551881 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-run-netns\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.551919 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/02375d8f-59c3-418c-98e5-901917ef970f-host-run-ovn-kubernetes\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.552441 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-ovnkube-script-lib\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.553319 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-ovnkube-config\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.553406 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/02375d8f-59c3-418c-98e5-901917ef970f-env-overrides\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.559183 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/02375d8f-59c3-418c-98e5-901917ef970f-ovn-node-metrics-cert\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.574401 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ll4t\" (UniqueName: \"kubernetes.io/projected/02375d8f-59c3-418c-98e5-901917ef970f-kube-api-access-2ll4t\") pod \"ovnkube-node-wd59t\" (UID: \"02375d8f-59c3-418c-98e5-901917ef970f\") " pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.585456 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.942695 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovnkube-controller/3.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.946999 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovn-acl-logging/0.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.947610 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-x6k66_af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/ovn-controller/0.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948019 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948056 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948066 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948076 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948085 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948095 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948104 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" exitCode=143 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948112 4979 generic.go:334] "Generic (PLEG): container finished" podID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" exitCode=143 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.948253 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950059 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950111 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950142 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950162 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950177 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950191 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950204 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950216 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950224 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950232 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950239 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950247 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950254 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950260 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950267 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950272 4979 scope.go:117] "RemoveContainer" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950276 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950539 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950566 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950576 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950663 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950705 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950715 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950724 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950733 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950742 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950750 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950773 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950796 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950808 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950818 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950827 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950836 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950846 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950855 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950864 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950874 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950883 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950897 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-x6k66" event={"ID":"af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26","Type":"ContainerDied","Data":"411c04f99f6258da1d397935c65aae70322c56cb7762d88705f89e71500e5dad"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950912 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950923 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950935 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950943 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950953 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.950989 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.951000 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.951010 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.951018 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.951059 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.951855 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/2.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.952572 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/1.log" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.952610 4979 generic.go:334] "Generic (PLEG): container finished" podID="12339390-77cf-4906-80c3-ddb729e42132" containerID="9a2f45253ea82811c08648bce91940b55482b206052274ef16b1d240ef1dda0c" exitCode=2 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.952741 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerDied","Data":"9a2f45253ea82811c08648bce91940b55482b206052274ef16b1d240ef1dda0c"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.952767 4979 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.953475 4979 scope.go:117] "RemoveContainer" containerID="9a2f45253ea82811c08648bce91940b55482b206052274ef16b1d240ef1dda0c" Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.956235 4979 generic.go:334] "Generic (PLEG): container finished" podID="02375d8f-59c3-418c-98e5-901917ef970f" containerID="b2e32da85c00c1ebf256a5c2aa5e60905a8bdd956bc2af269f9868c845d2ac68" exitCode=0 Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.956262 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerDied","Data":"b2e32da85c00c1ebf256a5c2aa5e60905a8bdd956bc2af269f9868c845d2ac68"} Dec 04 11:56:23 crc kubenswrapper[4979]: I1204 11:56:23.956282 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"1e2f002c2764644cee892da2b3b4bd342624af7e2fad7356b07c026e1e8e211e"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.011846 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.076680 4979 scope.go:117] "RemoveContainer" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.084472 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x6k66"] Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.089340 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-x6k66"] Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.112491 4979 scope.go:117] "RemoveContainer" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.129573 4979 scope.go:117] "RemoveContainer" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.145506 4979 scope.go:117] "RemoveContainer" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.160941 4979 scope.go:117] "RemoveContainer" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.175211 4979 scope.go:117] "RemoveContainer" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.192140 4979 scope.go:117] "RemoveContainer" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.207193 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26" path="/var/lib/kubelet/pods/af3c419d-ed2b-4fb7-bd28-5c6a5a8faa26/volumes" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.214662 4979 scope.go:117] "RemoveContainer" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.251586 4979 scope.go:117] "RemoveContainer" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.252229 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": container with ID starting with 6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f not found: ID does not exist" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.252273 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} err="failed to get container status \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": rpc error: code = NotFound desc = could not find container \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": container with ID starting with 6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.252312 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.252773 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": container with ID starting with 206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f not found: ID does not exist" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.252822 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} err="failed to get container status \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": rpc error: code = NotFound desc = could not find container \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": container with ID starting with 206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.252850 4979 scope.go:117] "RemoveContainer" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.253380 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": container with ID starting with 516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2 not found: ID does not exist" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.253409 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} err="failed to get container status \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": rpc error: code = NotFound desc = could not find container \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": container with ID starting with 516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.253426 4979 scope.go:117] "RemoveContainer" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.253659 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": container with ID starting with 7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790 not found: ID does not exist" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.253686 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} err="failed to get container status \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": rpc error: code = NotFound desc = could not find container \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": container with ID starting with 7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.253705 4979 scope.go:117] "RemoveContainer" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.253908 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": container with ID starting with 696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7 not found: ID does not exist" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.253928 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} err="failed to get container status \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": rpc error: code = NotFound desc = could not find container \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": container with ID starting with 696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.253941 4979 scope.go:117] "RemoveContainer" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.254131 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": container with ID starting with ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9 not found: ID does not exist" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.254151 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} err="failed to get container status \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": rpc error: code = NotFound desc = could not find container \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": container with ID starting with ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.254165 4979 scope.go:117] "RemoveContainer" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.254413 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": container with ID starting with b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae not found: ID does not exist" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.254431 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} err="failed to get container status \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": rpc error: code = NotFound desc = could not find container \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": container with ID starting with b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.254444 4979 scope.go:117] "RemoveContainer" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.254739 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": container with ID starting with d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df not found: ID does not exist" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.254759 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} err="failed to get container status \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": rpc error: code = NotFound desc = could not find container \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": container with ID starting with d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.254771 4979 scope.go:117] "RemoveContainer" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.255000 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": container with ID starting with a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c not found: ID does not exist" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255066 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} err="failed to get container status \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": rpc error: code = NotFound desc = could not find container \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": container with ID starting with a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255085 4979 scope.go:117] "RemoveContainer" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" Dec 04 11:56:24 crc kubenswrapper[4979]: E1204 11:56:24.255332 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": container with ID starting with 659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d not found: ID does not exist" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255370 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} err="failed to get container status \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": rpc error: code = NotFound desc = could not find container \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": container with ID starting with 659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255388 4979 scope.go:117] "RemoveContainer" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255617 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} err="failed to get container status \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": rpc error: code = NotFound desc = could not find container \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": container with ID starting with 6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255636 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255808 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} err="failed to get container status \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": rpc error: code = NotFound desc = could not find container \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": container with ID starting with 206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255828 4979 scope.go:117] "RemoveContainer" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255982 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} err="failed to get container status \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": rpc error: code = NotFound desc = could not find container \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": container with ID starting with 516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.255999 4979 scope.go:117] "RemoveContainer" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256180 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} err="failed to get container status \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": rpc error: code = NotFound desc = could not find container \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": container with ID starting with 7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256209 4979 scope.go:117] "RemoveContainer" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256437 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} err="failed to get container status \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": rpc error: code = NotFound desc = could not find container \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": container with ID starting with 696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256464 4979 scope.go:117] "RemoveContainer" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256670 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} err="failed to get container status \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": rpc error: code = NotFound desc = could not find container \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": container with ID starting with ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256690 4979 scope.go:117] "RemoveContainer" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256883 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} err="failed to get container status \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": rpc error: code = NotFound desc = could not find container \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": container with ID starting with b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.256903 4979 scope.go:117] "RemoveContainer" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257094 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} err="failed to get container status \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": rpc error: code = NotFound desc = could not find container \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": container with ID starting with d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257118 4979 scope.go:117] "RemoveContainer" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257394 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} err="failed to get container status \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": rpc error: code = NotFound desc = could not find container \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": container with ID starting with a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257421 4979 scope.go:117] "RemoveContainer" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257627 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} err="failed to get container status \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": rpc error: code = NotFound desc = could not find container \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": container with ID starting with 659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257648 4979 scope.go:117] "RemoveContainer" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257840 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} err="failed to get container status \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": rpc error: code = NotFound desc = could not find container \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": container with ID starting with 6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.257866 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258077 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} err="failed to get container status \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": rpc error: code = NotFound desc = could not find container \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": container with ID starting with 206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258100 4979 scope.go:117] "RemoveContainer" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258316 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} err="failed to get container status \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": rpc error: code = NotFound desc = could not find container \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": container with ID starting with 516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258336 4979 scope.go:117] "RemoveContainer" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258544 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} err="failed to get container status \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": rpc error: code = NotFound desc = could not find container \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": container with ID starting with 7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258570 4979 scope.go:117] "RemoveContainer" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258782 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} err="failed to get container status \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": rpc error: code = NotFound desc = could not find container \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": container with ID starting with 696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.258806 4979 scope.go:117] "RemoveContainer" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259038 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} err="failed to get container status \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": rpc error: code = NotFound desc = could not find container \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": container with ID starting with ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259066 4979 scope.go:117] "RemoveContainer" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259290 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} err="failed to get container status \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": rpc error: code = NotFound desc = could not find container \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": container with ID starting with b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259330 4979 scope.go:117] "RemoveContainer" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259666 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} err="failed to get container status \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": rpc error: code = NotFound desc = could not find container \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": container with ID starting with d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259698 4979 scope.go:117] "RemoveContainer" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.259982 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} err="failed to get container status \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": rpc error: code = NotFound desc = could not find container \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": container with ID starting with a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260008 4979 scope.go:117] "RemoveContainer" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260267 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} err="failed to get container status \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": rpc error: code = NotFound desc = could not find container \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": container with ID starting with 659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260364 4979 scope.go:117] "RemoveContainer" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260628 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} err="failed to get container status \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": rpc error: code = NotFound desc = could not find container \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": container with ID starting with 6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260657 4979 scope.go:117] "RemoveContainer" containerID="206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260876 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f"} err="failed to get container status \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": rpc error: code = NotFound desc = could not find container \"206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f\": container with ID starting with 206e481b64d50d287cc146ba1fcbd8105784c4d2eb9b6ec9b106e109bfb35a2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.260900 4979 scope.go:117] "RemoveContainer" containerID="516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261099 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2"} err="failed to get container status \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": rpc error: code = NotFound desc = could not find container \"516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2\": container with ID starting with 516b50de58cd22760e08b32e358fc7320da179791126a0f4fce8c1ba2e5035d2 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261122 4979 scope.go:117] "RemoveContainer" containerID="7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261345 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790"} err="failed to get container status \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": rpc error: code = NotFound desc = could not find container \"7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790\": container with ID starting with 7074ba32dde9a3c08a2f97c79af747c2c2d44285891c1506cbd4a72886987790 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261369 4979 scope.go:117] "RemoveContainer" containerID="696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261577 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7"} err="failed to get container status \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": rpc error: code = NotFound desc = could not find container \"696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7\": container with ID starting with 696ee2546908359533009f6b4acc599e0576c9213e71a64b43686a7d115997c7 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261607 4979 scope.go:117] "RemoveContainer" containerID="ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261930 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9"} err="failed to get container status \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": rpc error: code = NotFound desc = could not find container \"ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9\": container with ID starting with ec8ab4f60023fcd3fd9abbd08c850e0fcbfd2d2cf830051f8286b2f5231ef8a9 not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.261954 4979 scope.go:117] "RemoveContainer" containerID="b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.262241 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae"} err="failed to get container status \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": rpc error: code = NotFound desc = could not find container \"b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae\": container with ID starting with b1d071ec1ae7b2b889f6864a2213a72d223962e23de24a3dc120edb0678381ae not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.262269 4979 scope.go:117] "RemoveContainer" containerID="d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.262659 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df"} err="failed to get container status \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": rpc error: code = NotFound desc = could not find container \"d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df\": container with ID starting with d2429ccc7195d9476b07ed229a783227cd6cf1d8032a21535c29dc9b831593df not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.262682 4979 scope.go:117] "RemoveContainer" containerID="a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.264940 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c"} err="failed to get container status \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": rpc error: code = NotFound desc = could not find container \"a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c\": container with ID starting with a2a1bd47b51e8b22ebeecccf024995be42ebec7dbdd971470a5e28b96eaf989c not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.264965 4979 scope.go:117] "RemoveContainer" containerID="659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.265160 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d"} err="failed to get container status \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": rpc error: code = NotFound desc = could not find container \"659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d\": container with ID starting with 659a2be5f040414936e6d520d39afad5de20d16bdf57e665a227781580047d2d not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.265176 4979 scope.go:117] "RemoveContainer" containerID="6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.265731 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f"} err="failed to get container status \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": rpc error: code = NotFound desc = could not find container \"6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f\": container with ID starting with 6d5665fbf6b2cd10fcbf74d3652e30f875a03818ca96715871685317c4ae3e2f not found: ID does not exist" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.964276 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"ba08b0136d0f5df6a7ed9dc5a8d8a453bdb0dcf6e823d0a397d015d9cc442fb8"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.964741 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"38b796e1c665331a9daf802798d0305eaa9e0b706a285361e21b2543005b2b49"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.964764 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"f2915f2d72dbf4613a0560ac6a808e031029a38c83443afecff3bf966fc5e163"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.964779 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"6d36b3bc9d7ffefc0fc4f5d7f82cb617e8d62e5b4432971f11a36fe2f553e5f0"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.964793 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"dbe132effa4191979df909e8b5804a22f6f84acffba7775083cd7813b58d50a3"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.964810 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"9c0ed94ffb43a7ce9b67180377fa00a162bbdcfb60374566c4bf204196c2d000"} Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.968029 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/2.log" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.968436 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/1.log" Dec 04 11:56:24 crc kubenswrapper[4979]: I1204 11:56:24.968473 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-g4nk7" event={"ID":"12339390-77cf-4906-80c3-ddb729e42132","Type":"ContainerStarted","Data":"9b56d285732bf94a16cc4aadef012604fabd7b95f154f146afc9c463357d5625"} Dec 04 11:56:27 crc kubenswrapper[4979]: I1204 11:56:27.992824 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"fb04e276bdb96c0c7354d5c2552f4b9a963c148a8e8ef8b2883c90ba6d0f1931"} Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.008955 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" event={"ID":"02375d8f-59c3-418c-98e5-901917ef970f","Type":"ContainerStarted","Data":"1204de82e54fb49c5328850d84bfe03f55f2e51b3d73325de96b76632168cb7a"} Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.009379 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.009397 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.009494 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.032371 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.037131 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" podStartSLOduration=7.037110949 podStartE2EDuration="7.037110949s" podCreationTimestamp="2025-12-04 11:56:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:56:30.032778842 +0000 UTC m=+814.307074646" watchObservedRunningTime="2025-12-04 11:56:30.037110949 +0000 UTC m=+814.311406753" Dec 04 11:56:30 crc kubenswrapper[4979]: I1204 11:56:30.039634 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.325418 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-j692p"] Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.326396 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.329044 4979 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-tm2n7" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.329956 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.330713 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.330764 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.341038 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-j692p"] Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.351624 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-node-mnt\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.351677 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-crc-storage\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.351717 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb82b\" (UniqueName: \"kubernetes.io/projected/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-kube-api-access-zb82b\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.455196 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb82b\" (UniqueName: \"kubernetes.io/projected/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-kube-api-access-zb82b\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.455522 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-node-mnt\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.455615 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-crc-storage\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.455865 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-node-mnt\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.456188 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-crc-storage\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.472935 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb82b\" (UniqueName: \"kubernetes.io/projected/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-kube-api-access-zb82b\") pod \"crc-storage-crc-j692p\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: I1204 11:56:31.659879 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: E1204 11:56:31.690511 4979 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(19904a7afb4567f0e228854587d11a2aee9c783b2ca993de1484121cc26489be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 11:56:31 crc kubenswrapper[4979]: E1204 11:56:31.690594 4979 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(19904a7afb4567f0e228854587d11a2aee9c783b2ca993de1484121cc26489be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: E1204 11:56:31.690621 4979 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(19904a7afb4567f0e228854587d11a2aee9c783b2ca993de1484121cc26489be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:31 crc kubenswrapper[4979]: E1204 11:56:31.690668 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-j692p_crc-storage(8dc1c6f5-64eb-456b-b7ff-6f951b425df4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-j692p_crc-storage(8dc1c6f5-64eb-456b-b7ff-6f951b425df4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(19904a7afb4567f0e228854587d11a2aee9c783b2ca993de1484121cc26489be): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-j692p" podUID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" Dec 04 11:56:32 crc kubenswrapper[4979]: I1204 11:56:32.021403 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:32 crc kubenswrapper[4979]: I1204 11:56:32.021787 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:32 crc kubenswrapper[4979]: E1204 11:56:32.042343 4979 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(3b8e60aba3e47419390c87ef9a3458e197ac2b1ccbdefacde4f4282fc3fc44a0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 04 11:56:32 crc kubenswrapper[4979]: E1204 11:56:32.042680 4979 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(3b8e60aba3e47419390c87ef9a3458e197ac2b1ccbdefacde4f4282fc3fc44a0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:32 crc kubenswrapper[4979]: E1204 11:56:32.042701 4979 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(3b8e60aba3e47419390c87ef9a3458e197ac2b1ccbdefacde4f4282fc3fc44a0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:32 crc kubenswrapper[4979]: E1204 11:56:32.042755 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"crc-storage-crc-j692p_crc-storage(8dc1c6f5-64eb-456b-b7ff-6f951b425df4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"crc-storage-crc-j692p_crc-storage(8dc1c6f5-64eb-456b-b7ff-6f951b425df4)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_crc-storage-crc-j692p_crc-storage_8dc1c6f5-64eb-456b-b7ff-6f951b425df4_0(3b8e60aba3e47419390c87ef9a3458e197ac2b1ccbdefacde4f4282fc3fc44a0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="crc-storage/crc-storage-crc-j692p" podUID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" Dec 04 11:56:44 crc kubenswrapper[4979]: I1204 11:56:44.198753 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:44 crc kubenswrapper[4979]: I1204 11:56:44.201166 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:44 crc kubenswrapper[4979]: I1204 11:56:44.437156 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-j692p"] Dec 04 11:56:44 crc kubenswrapper[4979]: I1204 11:56:44.452231 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 11:56:45 crc kubenswrapper[4979]: I1204 11:56:45.124989 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-j692p" event={"ID":"8dc1c6f5-64eb-456b-b7ff-6f951b425df4","Type":"ContainerStarted","Data":"0de8f3e4eae57e2ee0bf4202a3ce8a63c929cf2a21ad1fbc35a13c61e384af62"} Dec 04 11:56:46 crc kubenswrapper[4979]: I1204 11:56:46.133976 4979 generic.go:334] "Generic (PLEG): container finished" podID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" containerID="e0f0996b0e0d34d127adb5109b2047117fcda81c7172b75f4198047276c964fe" exitCode=0 Dec 04 11:56:46 crc kubenswrapper[4979]: I1204 11:56:46.134070 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-j692p" event={"ID":"8dc1c6f5-64eb-456b-b7ff-6f951b425df4","Type":"ContainerDied","Data":"e0f0996b0e0d34d127adb5109b2047117fcda81c7172b75f4198047276c964fe"} Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.482091 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.674522 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb82b\" (UniqueName: \"kubernetes.io/projected/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-kube-api-access-zb82b\") pod \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.674611 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-crc-storage\") pod \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.674657 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-node-mnt\") pod \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\" (UID: \"8dc1c6f5-64eb-456b-b7ff-6f951b425df4\") " Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.674868 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "8dc1c6f5-64eb-456b-b7ff-6f951b425df4" (UID: "8dc1c6f5-64eb-456b-b7ff-6f951b425df4"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.675101 4979 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.680460 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-kube-api-access-zb82b" (OuterVolumeSpecName: "kube-api-access-zb82b") pod "8dc1c6f5-64eb-456b-b7ff-6f951b425df4" (UID: "8dc1c6f5-64eb-456b-b7ff-6f951b425df4"). InnerVolumeSpecName "kube-api-access-zb82b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.694597 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "8dc1c6f5-64eb-456b-b7ff-6f951b425df4" (UID: "8dc1c6f5-64eb-456b-b7ff-6f951b425df4"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.776379 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb82b\" (UniqueName: \"kubernetes.io/projected/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-kube-api-access-zb82b\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:47 crc kubenswrapper[4979]: I1204 11:56:47.776418 4979 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/8dc1c6f5-64eb-456b-b7ff-6f951b425df4-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 04 11:56:48 crc kubenswrapper[4979]: I1204 11:56:48.152000 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-j692p" event={"ID":"8dc1c6f5-64eb-456b-b7ff-6f951b425df4","Type":"ContainerDied","Data":"0de8f3e4eae57e2ee0bf4202a3ce8a63c929cf2a21ad1fbc35a13c61e384af62"} Dec 04 11:56:48 crc kubenswrapper[4979]: I1204 11:56:48.152112 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0de8f3e4eae57e2ee0bf4202a3ce8a63c929cf2a21ad1fbc35a13c61e384af62" Dec 04 11:56:48 crc kubenswrapper[4979]: I1204 11:56:48.152107 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-j692p" Dec 04 11:56:53 crc kubenswrapper[4979]: I1204 11:56:53.608664 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wd59t" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.505413 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff"] Dec 04 11:56:55 crc kubenswrapper[4979]: E1204 11:56:55.506001 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" containerName="storage" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.506015 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" containerName="storage" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.506114 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" containerName="storage" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.506935 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.509609 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.511281 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff"] Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.684493 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.684589 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.684789 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdmn8\" (UniqueName: \"kubernetes.io/projected/7ee51840-874e-4d0a-ac2c-5749ac25adec-kube-api-access-sdmn8\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.786655 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdmn8\" (UniqueName: \"kubernetes.io/projected/7ee51840-874e-4d0a-ac2c-5749ac25adec-kube-api-access-sdmn8\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.786879 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.786972 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.787560 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.787756 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.820838 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdmn8\" (UniqueName: \"kubernetes.io/projected/7ee51840-874e-4d0a-ac2c-5749ac25adec-kube-api-access-sdmn8\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:55 crc kubenswrapper[4979]: I1204 11:56:55.831763 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:56:56 crc kubenswrapper[4979]: I1204 11:56:56.163716 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff"] Dec 04 11:56:56 crc kubenswrapper[4979]: I1204 11:56:56.205940 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" event={"ID":"7ee51840-874e-4d0a-ac2c-5749ac25adec","Type":"ContainerStarted","Data":"974fe8de43149b0ca2509daba7eda8b9358d2aa6bc24f9d018c00ae15cf1091b"} Dec 04 11:56:56 crc kubenswrapper[4979]: I1204 11:56:56.500588 4979 scope.go:117] "RemoveContainer" containerID="40cc56d1ad8b90516cc64f8b039852d53c1174477ea606b5ac470d26a6975b93" Dec 04 11:56:57 crc kubenswrapper[4979]: I1204 11:56:57.211750 4979 generic.go:334] "Generic (PLEG): container finished" podID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerID="88876db4569fc77f9375027da24080aa022a843bc86bf7c5688213a7a95eeb50" exitCode=0 Dec 04 11:56:57 crc kubenswrapper[4979]: I1204 11:56:57.211810 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" event={"ID":"7ee51840-874e-4d0a-ac2c-5749ac25adec","Type":"ContainerDied","Data":"88876db4569fc77f9375027da24080aa022a843bc86bf7c5688213a7a95eeb50"} Dec 04 11:56:57 crc kubenswrapper[4979]: I1204 11:56:57.214771 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-g4nk7_12339390-77cf-4906-80c3-ddb729e42132/kube-multus/2.log" Dec 04 11:56:57 crc kubenswrapper[4979]: I1204 11:56:57.875827 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qtbpp"] Dec 04 11:56:57 crc kubenswrapper[4979]: I1204 11:56:57.884059 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:57 crc kubenswrapper[4979]: I1204 11:56:57.893773 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtbpp"] Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.013649 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-utilities\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.013703 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dh8x\" (UniqueName: \"kubernetes.io/projected/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-kube-api-access-5dh8x\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.013741 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-catalog-content\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.115316 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-utilities\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.115386 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dh8x\" (UniqueName: \"kubernetes.io/projected/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-kube-api-access-5dh8x\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.115434 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-catalog-content\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.115957 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-utilities\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.116011 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-catalog-content\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.133788 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dh8x\" (UniqueName: \"kubernetes.io/projected/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-kube-api-access-5dh8x\") pod \"redhat-operators-qtbpp\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.212858 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:56:58 crc kubenswrapper[4979]: I1204 11:56:58.468998 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qtbpp"] Dec 04 11:56:58 crc kubenswrapper[4979]: W1204 11:56:58.477188 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded3e2c6f_a5c3_41e6_94af_beb50dbf1aef.slice/crio-f50ff73dc8836d00c01597955540da08b3e4825992228bc12f45a38e0b29e925 WatchSource:0}: Error finding container f50ff73dc8836d00c01597955540da08b3e4825992228bc12f45a38e0b29e925: Status 404 returned error can't find the container with id f50ff73dc8836d00c01597955540da08b3e4825992228bc12f45a38e0b29e925 Dec 04 11:56:59 crc kubenswrapper[4979]: I1204 11:56:59.228163 4979 generic.go:334] "Generic (PLEG): container finished" podID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerID="debab1d2f0260dfb295045c29c499601d7ed2715a54219c6de2048e7580a17ec" exitCode=0 Dec 04 11:56:59 crc kubenswrapper[4979]: I1204 11:56:59.228210 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerDied","Data":"debab1d2f0260dfb295045c29c499601d7ed2715a54219c6de2048e7580a17ec"} Dec 04 11:56:59 crc kubenswrapper[4979]: I1204 11:56:59.228657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerStarted","Data":"f50ff73dc8836d00c01597955540da08b3e4825992228bc12f45a38e0b29e925"} Dec 04 11:56:59 crc kubenswrapper[4979]: I1204 11:56:59.232476 4979 generic.go:334] "Generic (PLEG): container finished" podID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerID="62d1c242e4cc57c0c77f41dc1f203cea552c1892e1ba01df53a696cb01580121" exitCode=0 Dec 04 11:56:59 crc kubenswrapper[4979]: I1204 11:56:59.232540 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" event={"ID":"7ee51840-874e-4d0a-ac2c-5749ac25adec","Type":"ContainerDied","Data":"62d1c242e4cc57c0c77f41dc1f203cea552c1892e1ba01df53a696cb01580121"} Dec 04 11:57:00 crc kubenswrapper[4979]: I1204 11:57:00.244291 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerStarted","Data":"5976a095fe0b8728fcb63b36a308f99297f1874de522c983a0fdb4ed9e9345c7"} Dec 04 11:57:00 crc kubenswrapper[4979]: I1204 11:57:00.248833 4979 generic.go:334] "Generic (PLEG): container finished" podID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerID="9af2f607564c7489d096d90865d89e370d1db973eb3275497bf1599cefa59bc5" exitCode=0 Dec 04 11:57:00 crc kubenswrapper[4979]: I1204 11:57:00.248888 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" event={"ID":"7ee51840-874e-4d0a-ac2c-5749ac25adec","Type":"ContainerDied","Data":"9af2f607564c7489d096d90865d89e370d1db973eb3275497bf1599cefa59bc5"} Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.254586 4979 generic.go:334] "Generic (PLEG): container finished" podID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerID="5976a095fe0b8728fcb63b36a308f99297f1874de522c983a0fdb4ed9e9345c7" exitCode=0 Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.255477 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerDied","Data":"5976a095fe0b8728fcb63b36a308f99297f1874de522c983a0fdb4ed9e9345c7"} Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.474556 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.590768 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdmn8\" (UniqueName: \"kubernetes.io/projected/7ee51840-874e-4d0a-ac2c-5749ac25adec-kube-api-access-sdmn8\") pod \"7ee51840-874e-4d0a-ac2c-5749ac25adec\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.591641 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-bundle\") pod \"7ee51840-874e-4d0a-ac2c-5749ac25adec\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.591708 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-util\") pod \"7ee51840-874e-4d0a-ac2c-5749ac25adec\" (UID: \"7ee51840-874e-4d0a-ac2c-5749ac25adec\") " Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.592402 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-bundle" (OuterVolumeSpecName: "bundle") pod "7ee51840-874e-4d0a-ac2c-5749ac25adec" (UID: "7ee51840-874e-4d0a-ac2c-5749ac25adec"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.598811 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ee51840-874e-4d0a-ac2c-5749ac25adec-kube-api-access-sdmn8" (OuterVolumeSpecName: "kube-api-access-sdmn8") pod "7ee51840-874e-4d0a-ac2c-5749ac25adec" (UID: "7ee51840-874e-4d0a-ac2c-5749ac25adec"). InnerVolumeSpecName "kube-api-access-sdmn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.621901 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-util" (OuterVolumeSpecName: "util") pod "7ee51840-874e-4d0a-ac2c-5749ac25adec" (UID: "7ee51840-874e-4d0a-ac2c-5749ac25adec"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.693481 4979 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.693514 4979 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7ee51840-874e-4d0a-ac2c-5749ac25adec-util\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:01 crc kubenswrapper[4979]: I1204 11:57:01.693525 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdmn8\" (UniqueName: \"kubernetes.io/projected/7ee51840-874e-4d0a-ac2c-5749ac25adec-kube-api-access-sdmn8\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:02 crc kubenswrapper[4979]: I1204 11:57:02.268983 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerStarted","Data":"ff116503cda858892121404c80f63b9c945c6d0ea835afc382b49126885a9804"} Dec 04 11:57:02 crc kubenswrapper[4979]: I1204 11:57:02.271591 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" event={"ID":"7ee51840-874e-4d0a-ac2c-5749ac25adec","Type":"ContainerDied","Data":"974fe8de43149b0ca2509daba7eda8b9358d2aa6bc24f9d018c00ae15cf1091b"} Dec 04 11:57:02 crc kubenswrapper[4979]: I1204 11:57:02.271621 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="974fe8de43149b0ca2509daba7eda8b9358d2aa6bc24f9d018c00ae15cf1091b" Dec 04 11:57:02 crc kubenswrapper[4979]: I1204 11:57:02.271656 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff" Dec 04 11:57:02 crc kubenswrapper[4979]: I1204 11:57:02.293664 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qtbpp" podStartSLOduration=2.830301169 podStartE2EDuration="5.29364664s" podCreationTimestamp="2025-12-04 11:56:57 +0000 UTC" firstStartedPulling="2025-12-04 11:56:59.230876861 +0000 UTC m=+843.505172695" lastFinishedPulling="2025-12-04 11:57:01.694222362 +0000 UTC m=+845.968518166" observedRunningTime="2025-12-04 11:57:02.290639268 +0000 UTC m=+846.564935082" watchObservedRunningTime="2025-12-04 11:57:02.29364664 +0000 UTC m=+846.567942444" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.775048 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8"] Dec 04 11:57:05 crc kubenswrapper[4979]: E1204 11:57:05.775545 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="extract" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.775556 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="extract" Dec 04 11:57:05 crc kubenswrapper[4979]: E1204 11:57:05.775572 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="util" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.775577 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="util" Dec 04 11:57:05 crc kubenswrapper[4979]: E1204 11:57:05.775595 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="pull" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.775601 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="pull" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.775706 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ee51840-874e-4d0a-ac2c-5749ac25adec" containerName="extract" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.776038 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.777639 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.777897 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-hdspp" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.779630 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.793621 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8"] Dec 04 11:57:05 crc kubenswrapper[4979]: I1204 11:57:05.962888 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmrwp\" (UniqueName: \"kubernetes.io/projected/42b69f76-cf35-4413-9e2a-12ae411b6d7d-kube-api-access-pmrwp\") pod \"nmstate-operator-5b5b58f5c8-2pbw8\" (UID: \"42b69f76-cf35-4413-9e2a-12ae411b6d7d\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" Dec 04 11:57:06 crc kubenswrapper[4979]: I1204 11:57:06.063602 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmrwp\" (UniqueName: \"kubernetes.io/projected/42b69f76-cf35-4413-9e2a-12ae411b6d7d-kube-api-access-pmrwp\") pod \"nmstate-operator-5b5b58f5c8-2pbw8\" (UID: \"42b69f76-cf35-4413-9e2a-12ae411b6d7d\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" Dec 04 11:57:06 crc kubenswrapper[4979]: I1204 11:57:06.088202 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmrwp\" (UniqueName: \"kubernetes.io/projected/42b69f76-cf35-4413-9e2a-12ae411b6d7d-kube-api-access-pmrwp\") pod \"nmstate-operator-5b5b58f5c8-2pbw8\" (UID: \"42b69f76-cf35-4413-9e2a-12ae411b6d7d\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" Dec 04 11:57:06 crc kubenswrapper[4979]: I1204 11:57:06.090074 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" Dec 04 11:57:06 crc kubenswrapper[4979]: I1204 11:57:06.285800 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8"] Dec 04 11:57:07 crc kubenswrapper[4979]: I1204 11:57:07.311503 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" event={"ID":"42b69f76-cf35-4413-9e2a-12ae411b6d7d","Type":"ContainerStarted","Data":"7e1b91c8289307b58e84b69cd592f55159ccd7b81121e40c55afef787e36f853"} Dec 04 11:57:08 crc kubenswrapper[4979]: I1204 11:57:08.213502 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:57:08 crc kubenswrapper[4979]: I1204 11:57:08.213583 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:57:08 crc kubenswrapper[4979]: I1204 11:57:08.268684 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:57:08 crc kubenswrapper[4979]: I1204 11:57:08.362451 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:57:10 crc kubenswrapper[4979]: I1204 11:57:10.331336 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" event={"ID":"42b69f76-cf35-4413-9e2a-12ae411b6d7d","Type":"ContainerStarted","Data":"9545272b6c4797e31f0e3182bee6c0e0620fe1978cf8d6527e86a571b80e4e59"} Dec 04 11:57:10 crc kubenswrapper[4979]: I1204 11:57:10.643787 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-2pbw8" podStartSLOduration=2.378462047 podStartE2EDuration="5.643752609s" podCreationTimestamp="2025-12-04 11:57:05 +0000 UTC" firstStartedPulling="2025-12-04 11:57:06.307876541 +0000 UTC m=+850.582172345" lastFinishedPulling="2025-12-04 11:57:09.573167103 +0000 UTC m=+853.847462907" observedRunningTime="2025-12-04 11:57:10.365263339 +0000 UTC m=+854.639559153" watchObservedRunningTime="2025-12-04 11:57:10.643752609 +0000 UTC m=+854.918048443" Dec 04 11:57:10 crc kubenswrapper[4979]: I1204 11:57:10.648723 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qtbpp"] Dec 04 11:57:10 crc kubenswrapper[4979]: I1204 11:57:10.649073 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qtbpp" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="registry-server" containerID="cri-o://ff116503cda858892121404c80f63b9c945c6d0ea835afc382b49126885a9804" gracePeriod=2 Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.343027 4979 generic.go:334] "Generic (PLEG): container finished" podID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerID="ff116503cda858892121404c80f63b9c945c6d0ea835afc382b49126885a9804" exitCode=0 Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.343116 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerDied","Data":"ff116503cda858892121404c80f63b9c945c6d0ea835afc382b49126885a9804"} Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.542767 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.745017 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-catalog-content\") pod \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.745216 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dh8x\" (UniqueName: \"kubernetes.io/projected/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-kube-api-access-5dh8x\") pod \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.745342 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-utilities\") pod \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\" (UID: \"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef\") " Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.747323 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-utilities" (OuterVolumeSpecName: "utilities") pod "ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" (UID: "ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.754391 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-kube-api-access-5dh8x" (OuterVolumeSpecName: "kube-api-access-5dh8x") pod "ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" (UID: "ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef"). InnerVolumeSpecName "kube-api-access-5dh8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.846319 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dh8x\" (UniqueName: \"kubernetes.io/projected/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-kube-api-access-5dh8x\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:11 crc kubenswrapper[4979]: I1204 11:57:11.846356 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.247560 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" (UID: "ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.250843 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.353702 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qtbpp" event={"ID":"ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef","Type":"ContainerDied","Data":"f50ff73dc8836d00c01597955540da08b3e4825992228bc12f45a38e0b29e925"} Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.353759 4979 scope.go:117] "RemoveContainer" containerID="ff116503cda858892121404c80f63b9c945c6d0ea835afc382b49126885a9804" Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.356504 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qtbpp" Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.377239 4979 scope.go:117] "RemoveContainer" containerID="5976a095fe0b8728fcb63b36a308f99297f1874de522c983a0fdb4ed9e9345c7" Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.399468 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qtbpp"] Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.406275 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qtbpp"] Dec 04 11:57:12 crc kubenswrapper[4979]: I1204 11:57:12.412619 4979 scope.go:117] "RemoveContainer" containerID="debab1d2f0260dfb295045c29c499601d7ed2715a54219c6de2048e7580a17ec" Dec 04 11:57:14 crc kubenswrapper[4979]: I1204 11:57:14.211831 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" path="/var/lib/kubelet/pods/ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef/volumes" Dec 04 11:57:15 crc kubenswrapper[4979]: I1204 11:57:15.998660 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn"] Dec 04 11:57:15 crc kubenswrapper[4979]: E1204 11:57:15.999238 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="extract-content" Dec 04 11:57:15 crc kubenswrapper[4979]: I1204 11:57:15.999254 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="extract-content" Dec 04 11:57:15 crc kubenswrapper[4979]: E1204 11:57:15.999270 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="extract-utilities" Dec 04 11:57:15 crc kubenswrapper[4979]: I1204 11:57:15.999281 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="extract-utilities" Dec 04 11:57:15 crc kubenswrapper[4979]: E1204 11:57:15.999342 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="registry-server" Dec 04 11:57:15 crc kubenswrapper[4979]: I1204 11:57:15.999353 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="registry-server" Dec 04 11:57:15 crc kubenswrapper[4979]: I1204 11:57:15.999476 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed3e2c6f-a5c3-41e6-94af-beb50dbf1aef" containerName="registry-server" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.000167 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.002928 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.003723 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.010039 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.010420 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vxfpb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.015026 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-fhws5"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.015950 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.019467 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.034569 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.121712 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.122324 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.127479 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.127570 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-bt62h" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.127779 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.149478 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.198482 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gb2mv\" (UniqueName: \"kubernetes.io/projected/515dbf63-c417-4af6-b570-a58812752c7f-kube-api-access-gb2mv\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.199861 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-nmstate-lock\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.199888 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t548w\" (UniqueName: \"kubernetes.io/projected/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-kube-api-access-t548w\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.199914 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8bfb\" (UniqueName: \"kubernetes.io/projected/6d6ff593-472b-4ae3-a015-a846b703daf8-kube-api-access-d8bfb\") pod \"nmstate-metrics-7f946cbc9-hwtsn\" (UID: \"6d6ff593-472b-4ae3-a015-a846b703daf8\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.199969 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-ovs-socket\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.199997 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-dbus-socket\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.200053 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.297347 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5b96db6c87-7rgdb"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.297984 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.300988 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-nmstate-lock\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301030 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t548w\" (UniqueName: \"kubernetes.io/projected/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-kube-api-access-t548w\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301057 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8bfb\" (UniqueName: \"kubernetes.io/projected/6d6ff593-472b-4ae3-a015-a846b703daf8-kube-api-access-d8bfb\") pod \"nmstate-metrics-7f946cbc9-hwtsn\" (UID: \"6d6ff593-472b-4ae3-a015-a846b703daf8\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301069 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-nmstate-lock\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301078 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-ovs-socket\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301394 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-ovs-socket\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301622 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-dbus-socket\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301664 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ac09b8ec-7526-45ed-9d0a-56c8416c635a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301693 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk7fl\" (UniqueName: \"kubernetes.io/projected/ac09b8ec-7526-45ed-9d0a-56c8416c635a-kube-api-access-pk7fl\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301938 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301957 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/515dbf63-c417-4af6-b570-a58812752c7f-dbus-socket\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.301976 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac09b8ec-7526-45ed-9d0a-56c8416c635a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.302040 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gb2mv\" (UniqueName: \"kubernetes.io/projected/515dbf63-c417-4af6-b570-a58812752c7f-kube-api-access-gb2mv\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: E1204 11:57:16.302226 4979 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 04 11:57:16 crc kubenswrapper[4979]: E1204 11:57:16.302270 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-tls-key-pair podName:e8a8bc91-603d-4d91-a54d-ec5c38845cb8 nodeName:}" failed. No retries permitted until 2025-12-04 11:57:16.802255472 +0000 UTC m=+861.076551276 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-7g9qj" (UID: "e8a8bc91-603d-4d91-a54d-ec5c38845cb8") : secret "openshift-nmstate-webhook" not found Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.326160 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t548w\" (UniqueName: \"kubernetes.io/projected/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-kube-api-access-t548w\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.332623 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gb2mv\" (UniqueName: \"kubernetes.io/projected/515dbf63-c417-4af6-b570-a58812752c7f-kube-api-access-gb2mv\") pod \"nmstate-handler-fhws5\" (UID: \"515dbf63-c417-4af6-b570-a58812752c7f\") " pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.343171 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8bfb\" (UniqueName: \"kubernetes.io/projected/6d6ff593-472b-4ae3-a015-a846b703daf8-kube-api-access-d8bfb\") pod \"nmstate-metrics-7f946cbc9-hwtsn\" (UID: \"6d6ff593-472b-4ae3-a015-a846b703daf8\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.363482 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b96db6c87-7rgdb"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.393664 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403224 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-oauth-config\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403262 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-service-ca\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403341 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ac09b8ec-7526-45ed-9d0a-56c8416c635a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403417 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk7fl\" (UniqueName: \"kubernetes.io/projected/ac09b8ec-7526-45ed-9d0a-56c8416c635a-kube-api-access-pk7fl\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403460 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-config\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403480 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-serving-cert\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403525 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac09b8ec-7526-45ed-9d0a-56c8416c635a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403553 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8xm8\" (UniqueName: \"kubernetes.io/projected/3b11fc15-4073-4f3f-bedd-98730a2578cb-kube-api-access-z8xm8\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.403874 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-trusted-ca-bundle\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.404004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-oauth-serving-cert\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.404344 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ac09b8ec-7526-45ed-9d0a-56c8416c635a-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.406968 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ac09b8ec-7526-45ed-9d0a-56c8416c635a-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.419843 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.420272 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk7fl\" (UniqueName: \"kubernetes.io/projected/ac09b8ec-7526-45ed-9d0a-56c8416c635a-kube-api-access-pk7fl\") pod \"nmstate-console-plugin-7fbb5f6569-5qhmm\" (UID: \"ac09b8ec-7526-45ed-9d0a-56c8416c635a\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.450568 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.508661 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-config\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.508703 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-serving-cert\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.508736 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8xm8\" (UniqueName: \"kubernetes.io/projected/3b11fc15-4073-4f3f-bedd-98730a2578cb-kube-api-access-z8xm8\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.508774 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-trusted-ca-bundle\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.508792 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-oauth-serving-cert\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.508820 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-oauth-config\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.509995 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-service-ca\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.510046 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-config\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.510455 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-trusted-ca-bundle\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.510554 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-oauth-serving-cert\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.510624 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3b11fc15-4073-4f3f-bedd-98730a2578cb-service-ca\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.512990 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-oauth-config\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.517732 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3b11fc15-4073-4f3f-bedd-98730a2578cb-console-serving-cert\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.526986 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8xm8\" (UniqueName: \"kubernetes.io/projected/3b11fc15-4073-4f3f-bedd-98730a2578cb-kube-api-access-z8xm8\") pod \"console-5b96db6c87-7rgdb\" (UID: \"3b11fc15-4073-4f3f-bedd-98730a2578cb\") " pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.609712 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.615807 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.658439 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.813320 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.817748 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5b96db6c87-7rgdb"] Dec 04 11:57:16 crc kubenswrapper[4979]: I1204 11:57:16.818921 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/e8a8bc91-603d-4d91-a54d-ec5c38845cb8-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-7g9qj\" (UID: \"e8a8bc91-603d-4d91-a54d-ec5c38845cb8\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:16 crc kubenswrapper[4979]: W1204 11:57:16.823497 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b11fc15_4073_4f3f_bedd_98730a2578cb.slice/crio-b8cc12f631f9202442a789c0b17c72f8b04f67fbbdd9839ad8613d767705ec13 WatchSource:0}: Error finding container b8cc12f631f9202442a789c0b17c72f8b04f67fbbdd9839ad8613d767705ec13: Status 404 returned error can't find the container with id b8cc12f631f9202442a789c0b17c72f8b04f67fbbdd9839ad8613d767705ec13 Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.009863 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.213225 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj"] Dec 04 11:57:17 crc kubenswrapper[4979]: W1204 11:57:17.218826 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode8a8bc91_603d_4d91_a54d_ec5c38845cb8.slice/crio-8c03c69c870c984c4478ecbfb041ae6d37385227eea20333bfe319cdca9b72ff WatchSource:0}: Error finding container 8c03c69c870c984c4478ecbfb041ae6d37385227eea20333bfe319cdca9b72ff: Status 404 returned error can't find the container with id 8c03c69c870c984c4478ecbfb041ae6d37385227eea20333bfe319cdca9b72ff Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.389878 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" event={"ID":"e8a8bc91-603d-4d91-a54d-ec5c38845cb8","Type":"ContainerStarted","Data":"8c03c69c870c984c4478ecbfb041ae6d37385227eea20333bfe319cdca9b72ff"} Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.391373 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" event={"ID":"6d6ff593-472b-4ae3-a015-a846b703daf8","Type":"ContainerStarted","Data":"ae15455fa9a2a5a1c78648c4429d0d7a201fe65b9aceeea49efdbc48b3d8f754"} Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.392548 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b96db6c87-7rgdb" event={"ID":"3b11fc15-4073-4f3f-bedd-98730a2578cb","Type":"ContainerStarted","Data":"b8cc12f631f9202442a789c0b17c72f8b04f67fbbdd9839ad8613d767705ec13"} Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.393832 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fhws5" event={"ID":"515dbf63-c417-4af6-b570-a58812752c7f","Type":"ContainerStarted","Data":"03ce78b6912a6adf49b2bb4411c465d258c1cb26ec5edec88b1f20524ee65148"} Dec 04 11:57:17 crc kubenswrapper[4979]: I1204 11:57:17.394915 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" event={"ID":"ac09b8ec-7526-45ed-9d0a-56c8416c635a","Type":"ContainerStarted","Data":"e9ca4a540f84b1eafd890a15fc87efec4bde7724ee2fc8247af61761eec2a0bb"} Dec 04 11:57:19 crc kubenswrapper[4979]: I1204 11:57:19.414005 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5b96db6c87-7rgdb" event={"ID":"3b11fc15-4073-4f3f-bedd-98730a2578cb","Type":"ContainerStarted","Data":"97abbf0d1f4f75620ead0735eaaee1d9c4a162895dec83b476e68f4e486eb1ee"} Dec 04 11:57:19 crc kubenswrapper[4979]: I1204 11:57:19.432330 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5b96db6c87-7rgdb" podStartSLOduration=3.432311043 podStartE2EDuration="3.432311043s" podCreationTimestamp="2025-12-04 11:57:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:57:19.428840588 +0000 UTC m=+863.703136412" watchObservedRunningTime="2025-12-04 11:57:19.432311043 +0000 UTC m=+863.706606867" Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.435579 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-fhws5" event={"ID":"515dbf63-c417-4af6-b570-a58812752c7f","Type":"ContainerStarted","Data":"10603c6c5845fbaf33b1d80d0625f99713ab8edcec932624267cc155d9cbf945"} Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.437568 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.441033 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" event={"ID":"e8a8bc91-603d-4d91-a54d-ec5c38845cb8","Type":"ContainerStarted","Data":"25abbe2ccad54a6492a49ff325ea23cac4e5991dff6ad47d712fb99f0753da85"} Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.441628 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.442943 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" event={"ID":"ac09b8ec-7526-45ed-9d0a-56c8416c635a","Type":"ContainerStarted","Data":"9ce738240d55bb7bfb718ff553663505559684feac61c4e0b83a4e9634e01747"} Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.444829 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" event={"ID":"6d6ff593-472b-4ae3-a015-a846b703daf8","Type":"ContainerStarted","Data":"60ef8537215e712505d017359abaf362df22c6ce76e3c13dbaf91758262c783c"} Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.452803 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-fhws5" podStartSLOduration=2.400120373 podStartE2EDuration="6.452788172s" podCreationTimestamp="2025-12-04 11:57:15 +0000 UTC" firstStartedPulling="2025-12-04 11:57:16.460962283 +0000 UTC m=+860.735258087" lastFinishedPulling="2025-12-04 11:57:20.513630082 +0000 UTC m=+864.787925886" observedRunningTime="2025-12-04 11:57:21.450693936 +0000 UTC m=+865.724989740" watchObservedRunningTime="2025-12-04 11:57:21.452788172 +0000 UTC m=+865.727083976" Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.483835 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5qhmm" podStartSLOduration=1.649123669 podStartE2EDuration="5.483811209s" podCreationTimestamp="2025-12-04 11:57:16 +0000 UTC" firstStartedPulling="2025-12-04 11:57:16.671353515 +0000 UTC m=+860.945649319" lastFinishedPulling="2025-12-04 11:57:20.506041055 +0000 UTC m=+864.780336859" observedRunningTime="2025-12-04 11:57:21.474229267 +0000 UTC m=+865.748525091" watchObservedRunningTime="2025-12-04 11:57:21.483811209 +0000 UTC m=+865.758107013" Dec 04 11:57:21 crc kubenswrapper[4979]: I1204 11:57:21.503151 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" podStartSLOduration=3.209180282 podStartE2EDuration="6.503132056s" podCreationTimestamp="2025-12-04 11:57:15 +0000 UTC" firstStartedPulling="2025-12-04 11:57:17.220549832 +0000 UTC m=+861.494845636" lastFinishedPulling="2025-12-04 11:57:20.514501576 +0000 UTC m=+864.788797410" observedRunningTime="2025-12-04 11:57:21.497443811 +0000 UTC m=+865.771739625" watchObservedRunningTime="2025-12-04 11:57:21.503132056 +0000 UTC m=+865.777427860" Dec 04 11:57:23 crc kubenswrapper[4979]: I1204 11:57:23.457476 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" event={"ID":"6d6ff593-472b-4ae3-a015-a846b703daf8","Type":"ContainerStarted","Data":"0cf89d55664c1760c2e32615dcefaca0fe032c86df316212ca6bc7fea2dc60a1"} Dec 04 11:57:26 crc kubenswrapper[4979]: I1204 11:57:26.447044 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-fhws5" Dec 04 11:57:26 crc kubenswrapper[4979]: I1204 11:57:26.464217 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-hwtsn" podStartSLOduration=5.464722517 podStartE2EDuration="11.464200086s" podCreationTimestamp="2025-12-04 11:57:15 +0000 UTC" firstStartedPulling="2025-12-04 11:57:16.633602794 +0000 UTC m=+860.907898598" lastFinishedPulling="2025-12-04 11:57:22.633080363 +0000 UTC m=+866.907376167" observedRunningTime="2025-12-04 11:57:23.479870393 +0000 UTC m=+867.754166187" watchObservedRunningTime="2025-12-04 11:57:26.464200086 +0000 UTC m=+870.738495900" Dec 04 11:57:26 crc kubenswrapper[4979]: I1204 11:57:26.616701 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:26 crc kubenswrapper[4979]: I1204 11:57:26.616799 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:26 crc kubenswrapper[4979]: I1204 11:57:26.623756 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:27 crc kubenswrapper[4979]: I1204 11:57:27.489479 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5b96db6c87-7rgdb" Dec 04 11:57:27 crc kubenswrapper[4979]: I1204 11:57:27.567592 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5wfhh"] Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.675230 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zm92h"] Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.676872 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.689835 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zm92h"] Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.847224 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-catalog-content\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.847315 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xr6sf\" (UniqueName: \"kubernetes.io/projected/655fd1c9-1e2f-4d92-8062-b8aed0716380-kube-api-access-xr6sf\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.847351 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-utilities\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.948606 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-catalog-content\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.948857 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xr6sf\" (UniqueName: \"kubernetes.io/projected/655fd1c9-1e2f-4d92-8062-b8aed0716380-kube-api-access-xr6sf\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.948885 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-utilities\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.949426 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-catalog-content\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.949457 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-utilities\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:33 crc kubenswrapper[4979]: I1204 11:57:33.973072 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xr6sf\" (UniqueName: \"kubernetes.io/projected/655fd1c9-1e2f-4d92-8062-b8aed0716380-kube-api-access-xr6sf\") pod \"redhat-marketplace-zm92h\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:34 crc kubenswrapper[4979]: I1204 11:57:34.040632 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:34 crc kubenswrapper[4979]: I1204 11:57:34.231508 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zm92h"] Dec 04 11:57:34 crc kubenswrapper[4979]: I1204 11:57:34.533269 4979 generic.go:334] "Generic (PLEG): container finished" podID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerID="a8cab8279aeadbcf40b52fa22967dbbd3fba4d2e8ee68b39b708035fd0722f16" exitCode=0 Dec 04 11:57:34 crc kubenswrapper[4979]: I1204 11:57:34.533462 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerDied","Data":"a8cab8279aeadbcf40b52fa22967dbbd3fba4d2e8ee68b39b708035fd0722f16"} Dec 04 11:57:34 crc kubenswrapper[4979]: I1204 11:57:34.533544 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerStarted","Data":"ca56be5340db2c037eafc5b0e47a8f57c88a4d016144bb88d3db199d99af8c01"} Dec 04 11:57:35 crc kubenswrapper[4979]: I1204 11:57:35.539712 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerStarted","Data":"bc377733d0bb381437552c7ad8bac9176d782f12624d203e38dc2b050933c284"} Dec 04 11:57:36 crc kubenswrapper[4979]: I1204 11:57:36.547150 4979 generic.go:334] "Generic (PLEG): container finished" podID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerID="bc377733d0bb381437552c7ad8bac9176d782f12624d203e38dc2b050933c284" exitCode=0 Dec 04 11:57:36 crc kubenswrapper[4979]: I1204 11:57:36.547195 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerDied","Data":"bc377733d0bb381437552c7ad8bac9176d782f12624d203e38dc2b050933c284"} Dec 04 11:57:36 crc kubenswrapper[4979]: I1204 11:57:36.547224 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerStarted","Data":"d29d305c1d0751e2aed2951e0aef209e5e2b5fb337ee2f67e8405871ef0dd775"} Dec 04 11:57:37 crc kubenswrapper[4979]: I1204 11:57:37.018372 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-7g9qj" Dec 04 11:57:37 crc kubenswrapper[4979]: I1204 11:57:37.037670 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zm92h" podStartSLOduration=2.628037751 podStartE2EDuration="4.03765078s" podCreationTimestamp="2025-12-04 11:57:33 +0000 UTC" firstStartedPulling="2025-12-04 11:57:34.535112735 +0000 UTC m=+878.809408539" lastFinishedPulling="2025-12-04 11:57:35.944725734 +0000 UTC m=+880.219021568" observedRunningTime="2025-12-04 11:57:36.571128929 +0000 UTC m=+880.845424743" watchObservedRunningTime="2025-12-04 11:57:37.03765078 +0000 UTC m=+881.311946604" Dec 04 11:57:44 crc kubenswrapper[4979]: I1204 11:57:44.043462 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:44 crc kubenswrapper[4979]: I1204 11:57:44.044064 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:44 crc kubenswrapper[4979]: I1204 11:57:44.103294 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:44 crc kubenswrapper[4979]: I1204 11:57:44.639316 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:44 crc kubenswrapper[4979]: I1204 11:57:44.676361 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zm92h"] Dec 04 11:57:46 crc kubenswrapper[4979]: I1204 11:57:46.607741 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zm92h" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="registry-server" containerID="cri-o://d29d305c1d0751e2aed2951e0aef209e5e2b5fb337ee2f67e8405871ef0dd775" gracePeriod=2 Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.614595 4979 generic.go:334] "Generic (PLEG): container finished" podID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerID="d29d305c1d0751e2aed2951e0aef209e5e2b5fb337ee2f67e8405871ef0dd775" exitCode=0 Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.614799 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerDied","Data":"d29d305c1d0751e2aed2951e0aef209e5e2b5fb337ee2f67e8405871ef0dd775"} Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.691861 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2l2xq"] Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.692857 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.705027 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2l2xq"] Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.749167 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-utilities\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.749226 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpcw7\" (UniqueName: \"kubernetes.io/projected/df4a8056-e9d3-43f2-b149-d761da531880-kube-api-access-dpcw7\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.749440 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-catalog-content\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.851382 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-catalog-content\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.851726 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-utilities\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.851747 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpcw7\" (UniqueName: \"kubernetes.io/projected/df4a8056-e9d3-43f2-b149-d761da531880-kube-api-access-dpcw7\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.852090 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-utilities\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.852093 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-catalog-content\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:47 crc kubenswrapper[4979]: I1204 11:57:47.869734 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpcw7\" (UniqueName: \"kubernetes.io/projected/df4a8056-e9d3-43f2-b149-d761da531880-kube-api-access-dpcw7\") pod \"certified-operators-2l2xq\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.011526 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.177647 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.265902 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-utilities\") pod \"655fd1c9-1e2f-4d92-8062-b8aed0716380\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.266183 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xr6sf\" (UniqueName: \"kubernetes.io/projected/655fd1c9-1e2f-4d92-8062-b8aed0716380-kube-api-access-xr6sf\") pod \"655fd1c9-1e2f-4d92-8062-b8aed0716380\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.266215 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-catalog-content\") pod \"655fd1c9-1e2f-4d92-8062-b8aed0716380\" (UID: \"655fd1c9-1e2f-4d92-8062-b8aed0716380\") " Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.267845 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-utilities" (OuterVolumeSpecName: "utilities") pod "655fd1c9-1e2f-4d92-8062-b8aed0716380" (UID: "655fd1c9-1e2f-4d92-8062-b8aed0716380"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.279973 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/655fd1c9-1e2f-4d92-8062-b8aed0716380-kube-api-access-xr6sf" (OuterVolumeSpecName: "kube-api-access-xr6sf") pod "655fd1c9-1e2f-4d92-8062-b8aed0716380" (UID: "655fd1c9-1e2f-4d92-8062-b8aed0716380"). InnerVolumeSpecName "kube-api-access-xr6sf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.292539 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "655fd1c9-1e2f-4d92-8062-b8aed0716380" (UID: "655fd1c9-1e2f-4d92-8062-b8aed0716380"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.368133 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.368168 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/655fd1c9-1e2f-4d92-8062-b8aed0716380-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.368183 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xr6sf\" (UniqueName: \"kubernetes.io/projected/655fd1c9-1e2f-4d92-8062-b8aed0716380-kube-api-access-xr6sf\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.483399 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2l2xq"] Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.621118 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2xq" event={"ID":"df4a8056-e9d3-43f2-b149-d761da531880","Type":"ContainerStarted","Data":"42cddcb2c2811329bca6cc2cf6e351e6fa56a637457c568d89512bce22f982d6"} Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.627742 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zm92h" event={"ID":"655fd1c9-1e2f-4d92-8062-b8aed0716380","Type":"ContainerDied","Data":"ca56be5340db2c037eafc5b0e47a8f57c88a4d016144bb88d3db199d99af8c01"} Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.627798 4979 scope.go:117] "RemoveContainer" containerID="d29d305c1d0751e2aed2951e0aef209e5e2b5fb337ee2f67e8405871ef0dd775" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.627801 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zm92h" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.644502 4979 scope.go:117] "RemoveContainer" containerID="bc377733d0bb381437552c7ad8bac9176d782f12624d203e38dc2b050933c284" Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.661351 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zm92h"] Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.665071 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zm92h"] Dec 04 11:57:48 crc kubenswrapper[4979]: I1204 11:57:48.693905 4979 scope.go:117] "RemoveContainer" containerID="a8cab8279aeadbcf40b52fa22967dbbd3fba4d2e8ee68b39b708035fd0722f16" Dec 04 11:57:49 crc kubenswrapper[4979]: I1204 11:57:49.679162 4979 generic.go:334] "Generic (PLEG): container finished" podID="df4a8056-e9d3-43f2-b149-d761da531880" containerID="8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c" exitCode=0 Dec 04 11:57:49 crc kubenswrapper[4979]: I1204 11:57:49.679846 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2xq" event={"ID":"df4a8056-e9d3-43f2-b149-d761da531880","Type":"ContainerDied","Data":"8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c"} Dec 04 11:57:50 crc kubenswrapper[4979]: I1204 11:57:50.208424 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" path="/var/lib/kubelet/pods/655fd1c9-1e2f-4d92-8062-b8aed0716380/volumes" Dec 04 11:57:51 crc kubenswrapper[4979]: I1204 11:57:51.692155 4979 generic.go:334] "Generic (PLEG): container finished" podID="df4a8056-e9d3-43f2-b149-d761da531880" containerID="f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4" exitCode=0 Dec 04 11:57:51 crc kubenswrapper[4979]: I1204 11:57:51.692211 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2xq" event={"ID":"df4a8056-e9d3-43f2-b149-d761da531880","Type":"ContainerDied","Data":"f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4"} Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.616000 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-5wfhh" podUID="82926c75-1f3c-4230-b981-f341f4e17ee2" containerName="console" containerID="cri-o://2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856" gracePeriod=15 Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.701027 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2xq" event={"ID":"df4a8056-e9d3-43f2-b149-d761da531880","Type":"ContainerStarted","Data":"4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382"} Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.706618 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7"] Dec 04 11:57:52 crc kubenswrapper[4979]: E1204 11:57:52.706863 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="extract-utilities" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.706881 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="extract-utilities" Dec 04 11:57:52 crc kubenswrapper[4979]: E1204 11:57:52.706900 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="registry-server" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.706906 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="registry-server" Dec 04 11:57:52 crc kubenswrapper[4979]: E1204 11:57:52.706916 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="extract-content" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.706922 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="extract-content" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.707022 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="655fd1c9-1e2f-4d92-8062-b8aed0716380" containerName="registry-server" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.707727 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.709398 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.718748 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7"] Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.728087 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2l2xq" podStartSLOduration=3.319870548 podStartE2EDuration="5.728072s" podCreationTimestamp="2025-12-04 11:57:47 +0000 UTC" firstStartedPulling="2025-12-04 11:57:49.681927169 +0000 UTC m=+893.956222973" lastFinishedPulling="2025-12-04 11:57:52.090128621 +0000 UTC m=+896.364424425" observedRunningTime="2025-12-04 11:57:52.727639098 +0000 UTC m=+897.001934902" watchObservedRunningTime="2025-12-04 11:57:52.728072 +0000 UTC m=+897.002367804" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.821899 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.821957 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.822013 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frhz8\" (UniqueName: \"kubernetes.io/projected/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-kube-api-access-frhz8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.924110 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.924170 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.924211 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frhz8\" (UniqueName: \"kubernetes.io/projected/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-kube-api-access-frhz8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.924668 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.925206 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.950764 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frhz8\" (UniqueName: \"kubernetes.io/projected/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-kube-api-access-frhz8\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.990834 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5wfhh_82926c75-1f3c-4230-b981-f341f4e17ee2/console/0.log" Dec 04 11:57:52 crc kubenswrapper[4979]: I1204 11:57:52.990921 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.022168 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.126910 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-serving-cert\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.126974 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-service-ca\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.127019 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbsr5\" (UniqueName: \"kubernetes.io/projected/82926c75-1f3c-4230-b981-f341f4e17ee2-kube-api-access-bbsr5\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.127074 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-oauth-serving-cert\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.127092 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-console-config\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.128011 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-oauth-config\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.128039 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-trusted-ca-bundle\") pod \"82926c75-1f3c-4230-b981-f341f4e17ee2\" (UID: \"82926c75-1f3c-4230-b981-f341f4e17ee2\") " Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.128628 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-service-ca" (OuterVolumeSpecName: "service-ca") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.128875 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.129378 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.129906 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-console-config" (OuterVolumeSpecName: "console-config") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.131205 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.131561 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82926c75-1f3c-4230-b981-f341f4e17ee2-kube-api-access-bbsr5" (OuterVolumeSpecName: "kube-api-access-bbsr5") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "kube-api-access-bbsr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.132756 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "82926c75-1f3c-4230-b981-f341f4e17ee2" (UID: "82926c75-1f3c-4230-b981-f341f4e17ee2"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.213674 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7"] Dec 04 11:57:53 crc kubenswrapper[4979]: W1204 11:57:53.214540 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c193f7_63b8_4a74_bfd3_29ac10d5afd6.slice/crio-f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3 WatchSource:0}: Error finding container f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3: Status 404 returned error can't find the container with id f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3 Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228883 4979 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228925 4979 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-console-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228937 4979 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228948 4979 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228960 4979 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/82926c75-1f3c-4230-b981-f341f4e17ee2-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228971 4979 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/82926c75-1f3c-4230-b981-f341f4e17ee2-service-ca\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.228982 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbsr5\" (UniqueName: \"kubernetes.io/projected/82926c75-1f3c-4230-b981-f341f4e17ee2-kube-api-access-bbsr5\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.707731 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-5wfhh_82926c75-1f3c-4230-b981-f341f4e17ee2/console/0.log" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.708032 4979 generic.go:334] "Generic (PLEG): container finished" podID="82926c75-1f3c-4230-b981-f341f4e17ee2" containerID="2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856" exitCode=2 Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.708106 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5wfhh" event={"ID":"82926c75-1f3c-4230-b981-f341f4e17ee2","Type":"ContainerDied","Data":"2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856"} Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.708145 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-5wfhh" event={"ID":"82926c75-1f3c-4230-b981-f341f4e17ee2","Type":"ContainerDied","Data":"0f0a2eb21f11096a22e4f334377e853462c11e04474f9bf7e1a383103c687db8"} Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.708156 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-5wfhh" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.708171 4979 scope.go:117] "RemoveContainer" containerID="2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.711711 4979 generic.go:334] "Generic (PLEG): container finished" podID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerID="65fc540c9f0594165e62bbb24d4cbe00b7f3778a7af81659e6397321ce912448" exitCode=0 Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.712817 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" event={"ID":"08c193f7-63b8-4a74-bfd3-29ac10d5afd6","Type":"ContainerDied","Data":"65fc540c9f0594165e62bbb24d4cbe00b7f3778a7af81659e6397321ce912448"} Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.712849 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" event={"ID":"08c193f7-63b8-4a74-bfd3-29ac10d5afd6","Type":"ContainerStarted","Data":"f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3"} Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.728946 4979 scope.go:117] "RemoveContainer" containerID="2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856" Dec 04 11:57:53 crc kubenswrapper[4979]: E1204 11:57:53.729442 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856\": container with ID starting with 2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856 not found: ID does not exist" containerID="2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.729479 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856"} err="failed to get container status \"2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856\": rpc error: code = NotFound desc = could not find container \"2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856\": container with ID starting with 2892e266a05804b722a8a910d1a2ca2f8e824ee8d67dfbad99796e9016dc8856 not found: ID does not exist" Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.740991 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-5wfhh"] Dec 04 11:57:53 crc kubenswrapper[4979]: I1204 11:57:53.745294 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-5wfhh"] Dec 04 11:57:54 crc kubenswrapper[4979]: I1204 11:57:54.206994 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82926c75-1f3c-4230-b981-f341f4e17ee2" path="/var/lib/kubelet/pods/82926c75-1f3c-4230-b981-f341f4e17ee2/volumes" Dec 04 11:57:55 crc kubenswrapper[4979]: I1204 11:57:55.730246 4979 generic.go:334] "Generic (PLEG): container finished" podID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerID="9eaac7539ec7439da1d4d373b0e75a0e2cc4b1060842d317e2036201d57854a2" exitCode=0 Dec 04 11:57:55 crc kubenswrapper[4979]: I1204 11:57:55.730365 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" event={"ID":"08c193f7-63b8-4a74-bfd3-29ac10d5afd6","Type":"ContainerDied","Data":"9eaac7539ec7439da1d4d373b0e75a0e2cc4b1060842d317e2036201d57854a2"} Dec 04 11:57:56 crc kubenswrapper[4979]: W1204 11:57:56.220807 4979 container.go:586] Failed to update stats for container "/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c193f7_63b8_4a74_bfd3_29ac10d5afd6.slice/crio-f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3": error while statting cgroup v2: [openat2 /sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c193f7_63b8_4a74_bfd3_29ac10d5afd6.slice/crio-f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3/memory.swap.current: no such device], continuing to push stats Dec 04 11:57:56 crc kubenswrapper[4979]: I1204 11:57:56.741906 4979 generic.go:334] "Generic (PLEG): container finished" podID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerID="e1be6caf9bb8c75954cf53ee4497b3d971d61821f0c9aef656f71145a1a7d278" exitCode=0 Dec 04 11:57:56 crc kubenswrapper[4979]: I1204 11:57:56.742082 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" event={"ID":"08c193f7-63b8-4a74-bfd3-29ac10d5afd6","Type":"ContainerDied","Data":"e1be6caf9bb8c75954cf53ee4497b3d971d61821f0c9aef656f71145a1a7d278"} Dec 04 11:57:57 crc kubenswrapper[4979]: E1204 11:57:57.282963 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08c193f7_63b8_4a74_bfd3_29ac10d5afd6.slice/crio-f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3\": RecentStats: unable to find data in memory cache]" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.012159 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.012207 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.039680 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.040750 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.040819 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.058395 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.195942 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-bundle\") pod \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.196008 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frhz8\" (UniqueName: \"kubernetes.io/projected/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-kube-api-access-frhz8\") pod \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.196107 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-util\") pod \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\" (UID: \"08c193f7-63b8-4a74-bfd3-29ac10d5afd6\") " Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.197349 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-bundle" (OuterVolumeSpecName: "bundle") pod "08c193f7-63b8-4a74-bfd3-29ac10d5afd6" (UID: "08c193f7-63b8-4a74-bfd3-29ac10d5afd6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.201785 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-kube-api-access-frhz8" (OuterVolumeSpecName: "kube-api-access-frhz8") pod "08c193f7-63b8-4a74-bfd3-29ac10d5afd6" (UID: "08c193f7-63b8-4a74-bfd3-29ac10d5afd6"). InnerVolumeSpecName "kube-api-access-frhz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.223578 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-util" (OuterVolumeSpecName: "util") pod "08c193f7-63b8-4a74-bfd3-29ac10d5afd6" (UID: "08c193f7-63b8-4a74-bfd3-29ac10d5afd6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.298772 4979 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.298972 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frhz8\" (UniqueName: \"kubernetes.io/projected/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-kube-api-access-frhz8\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.299104 4979 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/08c193f7-63b8-4a74-bfd3-29ac10d5afd6-util\") on node \"crc\" DevicePath \"\"" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.759412 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" event={"ID":"08c193f7-63b8-4a74-bfd3-29ac10d5afd6","Type":"ContainerDied","Data":"f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3"} Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.759492 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9c5083e3289c488dae00bc7f28208904aa151be391bf6f427ace4a5a11c45a3" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.759502 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7" Dec 04 11:57:58 crc kubenswrapper[4979]: I1204 11:57:58.810021 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:58:00 crc kubenswrapper[4979]: I1204 11:58:00.668811 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2l2xq"] Dec 04 11:58:00 crc kubenswrapper[4979]: I1204 11:58:00.771211 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2l2xq" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="registry-server" containerID="cri-o://4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382" gracePeriod=2 Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.083234 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.234110 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-utilities\") pod \"df4a8056-e9d3-43f2-b149-d761da531880\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.234174 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-catalog-content\") pod \"df4a8056-e9d3-43f2-b149-d761da531880\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.234252 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpcw7\" (UniqueName: \"kubernetes.io/projected/df4a8056-e9d3-43f2-b149-d761da531880-kube-api-access-dpcw7\") pod \"df4a8056-e9d3-43f2-b149-d761da531880\" (UID: \"df4a8056-e9d3-43f2-b149-d761da531880\") " Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.234907 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-utilities" (OuterVolumeSpecName: "utilities") pod "df4a8056-e9d3-43f2-b149-d761da531880" (UID: "df4a8056-e9d3-43f2-b149-d761da531880"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.240147 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df4a8056-e9d3-43f2-b149-d761da531880-kube-api-access-dpcw7" (OuterVolumeSpecName: "kube-api-access-dpcw7") pod "df4a8056-e9d3-43f2-b149-d761da531880" (UID: "df4a8056-e9d3-43f2-b149-d761da531880"). InnerVolumeSpecName "kube-api-access-dpcw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.309018 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df4a8056-e9d3-43f2-b149-d761da531880" (UID: "df4a8056-e9d3-43f2-b149-d761da531880"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.335928 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpcw7\" (UniqueName: \"kubernetes.io/projected/df4a8056-e9d3-43f2-b149-d761da531880-kube-api-access-dpcw7\") on node \"crc\" DevicePath \"\"" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.335965 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.335977 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df4a8056-e9d3-43f2-b149-d761da531880-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.777482 4979 generic.go:334] "Generic (PLEG): container finished" podID="df4a8056-e9d3-43f2-b149-d761da531880" containerID="4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382" exitCode=0 Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.777536 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2l2xq" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.777553 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2xq" event={"ID":"df4a8056-e9d3-43f2-b149-d761da531880","Type":"ContainerDied","Data":"4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382"} Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.777934 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2l2xq" event={"ID":"df4a8056-e9d3-43f2-b149-d761da531880","Type":"ContainerDied","Data":"42cddcb2c2811329bca6cc2cf6e351e6fa56a637457c568d89512bce22f982d6"} Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.777953 4979 scope.go:117] "RemoveContainer" containerID="4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.794662 4979 scope.go:117] "RemoveContainer" containerID="f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.810716 4979 scope.go:117] "RemoveContainer" containerID="8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.818192 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2l2xq"] Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.824468 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2l2xq"] Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.831933 4979 scope.go:117] "RemoveContainer" containerID="4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382" Dec 04 11:58:01 crc kubenswrapper[4979]: E1204 11:58:01.832461 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382\": container with ID starting with 4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382 not found: ID does not exist" containerID="4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.832521 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382"} err="failed to get container status \"4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382\": rpc error: code = NotFound desc = could not find container \"4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382\": container with ID starting with 4f547c4224652a8004741796ac5a51657b65dd42d63c085a69d061fab42b1382 not found: ID does not exist" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.832558 4979 scope.go:117] "RemoveContainer" containerID="f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4" Dec 04 11:58:01 crc kubenswrapper[4979]: E1204 11:58:01.832890 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4\": container with ID starting with f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4 not found: ID does not exist" containerID="f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.832943 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4"} err="failed to get container status \"f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4\": rpc error: code = NotFound desc = could not find container \"f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4\": container with ID starting with f2de8dc1978199ec99acf49df23571d30522bd5c7344b7b517af26da5edb2ae4 not found: ID does not exist" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.832975 4979 scope.go:117] "RemoveContainer" containerID="8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c" Dec 04 11:58:01 crc kubenswrapper[4979]: E1204 11:58:01.835414 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c\": container with ID starting with 8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c not found: ID does not exist" containerID="8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c" Dec 04 11:58:01 crc kubenswrapper[4979]: I1204 11:58:01.835444 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c"} err="failed to get container status \"8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c\": rpc error: code = NotFound desc = could not find container \"8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c\": container with ID starting with 8416d278ad3cd85152a0df33dba1f57f3941c43085e734f0853298386eaff60c not found: ID does not exist" Dec 04 11:58:02 crc kubenswrapper[4979]: I1204 11:58:02.206254 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df4a8056-e9d3-43f2-b149-d761da531880" path="/var/lib/kubelet/pods/df4a8056-e9d3-43f2-b149-d761da531880/volumes" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422184 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh"] Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422399 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="extract-content" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422411 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="extract-content" Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422426 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82926c75-1f3c-4230-b981-f341f4e17ee2" containerName="console" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422431 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="82926c75-1f3c-4230-b981-f341f4e17ee2" containerName="console" Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422440 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="registry-server" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422446 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="registry-server" Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422451 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="util" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422458 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="util" Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422465 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="pull" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422470 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="pull" Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422477 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="extract" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422482 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="extract" Dec 04 11:58:06 crc kubenswrapper[4979]: E1204 11:58:06.422489 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="extract-utilities" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422494 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="extract-utilities" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422592 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="08c193f7-63b8-4a74-bfd3-29ac10d5afd6" containerName="extract" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422603 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="82926c75-1f3c-4230-b981-f341f4e17ee2" containerName="console" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422614 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="df4a8056-e9d3-43f2-b149-d761da531880" containerName="registry-server" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.422994 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.426217 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.426738 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.426908 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.427034 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.427060 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8nqpp" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.443548 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh"] Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.604995 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06d63892-a58b-4921-9163-acdb6808729d-webhook-cert\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.605049 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s75z7\" (UniqueName: \"kubernetes.io/projected/06d63892-a58b-4921-9163-acdb6808729d-kube-api-access-s75z7\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.605176 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06d63892-a58b-4921-9163-acdb6808729d-apiservice-cert\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.706862 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06d63892-a58b-4921-9163-acdb6808729d-apiservice-cert\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.706985 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06d63892-a58b-4921-9163-acdb6808729d-webhook-cert\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.707022 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s75z7\" (UniqueName: \"kubernetes.io/projected/06d63892-a58b-4921-9163-acdb6808729d-kube-api-access-s75z7\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.713948 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06d63892-a58b-4921-9163-acdb6808729d-apiservice-cert\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.714578 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06d63892-a58b-4921-9163-acdb6808729d-webhook-cert\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.728941 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s75z7\" (UniqueName: \"kubernetes.io/projected/06d63892-a58b-4921-9163-acdb6808729d-kube-api-access-s75z7\") pod \"metallb-operator-controller-manager-7d59d98bc-t8mjh\" (UID: \"06d63892-a58b-4921-9163-acdb6808729d\") " pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.744739 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.771317 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7"] Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.772728 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.776791 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-c7vld" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.777045 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.777226 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.781742 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7"] Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.909219 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e81d70c-96a3-4a8d-b817-534a20d99799-webhook-cert\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.909253 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hv4c\" (UniqueName: \"kubernetes.io/projected/6e81d70c-96a3-4a8d-b817-534a20d99799-kube-api-access-2hv4c\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:06 crc kubenswrapper[4979]: I1204 11:58:06.909288 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e81d70c-96a3-4a8d-b817-534a20d99799-apiservice-cert\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.006846 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh"] Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.009950 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e81d70c-96a3-4a8d-b817-534a20d99799-webhook-cert\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.009989 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hv4c\" (UniqueName: \"kubernetes.io/projected/6e81d70c-96a3-4a8d-b817-534a20d99799-kube-api-access-2hv4c\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.010020 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e81d70c-96a3-4a8d-b817-534a20d99799-apiservice-cert\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.018245 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/6e81d70c-96a3-4a8d-b817-534a20d99799-webhook-cert\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.025175 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/6e81d70c-96a3-4a8d-b817-534a20d99799-apiservice-cert\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.035404 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hv4c\" (UniqueName: \"kubernetes.io/projected/6e81d70c-96a3-4a8d-b817-534a20d99799-kube-api-access-2hv4c\") pod \"metallb-operator-webhook-server-86498874f8-zlxh7\" (UID: \"6e81d70c-96a3-4a8d-b817-534a20d99799\") " pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.110872 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.531841 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7"] Dec 04 11:58:07 crc kubenswrapper[4979]: W1204 11:58:07.538136 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e81d70c_96a3_4a8d_b817_534a20d99799.slice/crio-b336ddee31f10eb2bc9d2350ee1e6f8bc6c0642b70badd94e43a48306ed382ae WatchSource:0}: Error finding container b336ddee31f10eb2bc9d2350ee1e6f8bc6c0642b70badd94e43a48306ed382ae: Status 404 returned error can't find the container with id b336ddee31f10eb2bc9d2350ee1e6f8bc6c0642b70badd94e43a48306ed382ae Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.834946 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" event={"ID":"06d63892-a58b-4921-9163-acdb6808729d","Type":"ContainerStarted","Data":"d336ada848d3f3aa37b120e5c59b6404bac8fcec73112328d2e7d10efb7e187c"} Dec 04 11:58:07 crc kubenswrapper[4979]: I1204 11:58:07.836836 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" event={"ID":"6e81d70c-96a3-4a8d-b817-534a20d99799","Type":"ContainerStarted","Data":"b336ddee31f10eb2bc9d2350ee1e6f8bc6c0642b70badd94e43a48306ed382ae"} Dec 04 11:58:10 crc kubenswrapper[4979]: I1204 11:58:10.864744 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" event={"ID":"06d63892-a58b-4921-9163-acdb6808729d","Type":"ContainerStarted","Data":"457b24d65eb4b1e909645fdfa4f91ed07fd487597b1db5fe217d01a855b89f15"} Dec 04 11:58:10 crc kubenswrapper[4979]: I1204 11:58:10.865167 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:10 crc kubenswrapper[4979]: I1204 11:58:10.889028 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" podStartSLOduration=1.9958314339999998 podStartE2EDuration="4.889007404s" podCreationTimestamp="2025-12-04 11:58:06 +0000 UTC" firstStartedPulling="2025-12-04 11:58:07.026853328 +0000 UTC m=+911.301149132" lastFinishedPulling="2025-12-04 11:58:09.920029298 +0000 UTC m=+914.194325102" observedRunningTime="2025-12-04 11:58:10.884704006 +0000 UTC m=+915.158999840" watchObservedRunningTime="2025-12-04 11:58:10.889007404 +0000 UTC m=+915.163303208" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.388024 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4v6w8"] Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.399709 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4v6w8"] Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.399805 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.480577 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-utilities\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.480676 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52xmp\" (UniqueName: \"kubernetes.io/projected/40453e49-262b-47eb-ba13-a0de8531e17b-kube-api-access-52xmp\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.480715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-catalog-content\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.584112 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-utilities\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.584192 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52xmp\" (UniqueName: \"kubernetes.io/projected/40453e49-262b-47eb-ba13-a0de8531e17b-kube-api-access-52xmp\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.584226 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-catalog-content\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.584790 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-catalog-content\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.584793 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-utilities\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.604330 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52xmp\" (UniqueName: \"kubernetes.io/projected/40453e49-262b-47eb-ba13-a0de8531e17b-kube-api-access-52xmp\") pod \"community-operators-4v6w8\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:11 crc kubenswrapper[4979]: I1204 11:58:11.724663 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.482959 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4v6w8"] Dec 04 11:58:13 crc kubenswrapper[4979]: W1204 11:58:13.492030 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40453e49_262b_47eb_ba13_a0de8531e17b.slice/crio-1e3a0e3e7e5af097bfd9a8cd1a6fd37688e37dffedfcee7909dad831e9c9edb4 WatchSource:0}: Error finding container 1e3a0e3e7e5af097bfd9a8cd1a6fd37688e37dffedfcee7909dad831e9c9edb4: Status 404 returned error can't find the container with id 1e3a0e3e7e5af097bfd9a8cd1a6fd37688e37dffedfcee7909dad831e9c9edb4 Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.886338 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" event={"ID":"6e81d70c-96a3-4a8d-b817-534a20d99799","Type":"ContainerStarted","Data":"80ac66943d2f27048061383ba61fe6fd0a9ba5b86fe675b8d6365d09a660c7bb"} Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.886488 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.888163 4979 generic.go:334] "Generic (PLEG): container finished" podID="40453e49-262b-47eb-ba13-a0de8531e17b" containerID="d2afe584aae1109d0aca0e2f22e9f5dcd0bb369c5b349fa801070f28ec54234b" exitCode=0 Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.888223 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4v6w8" event={"ID":"40453e49-262b-47eb-ba13-a0de8531e17b","Type":"ContainerDied","Data":"d2afe584aae1109d0aca0e2f22e9f5dcd0bb369c5b349fa801070f28ec54234b"} Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.888246 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4v6w8" event={"ID":"40453e49-262b-47eb-ba13-a0de8531e17b","Type":"ContainerStarted","Data":"1e3a0e3e7e5af097bfd9a8cd1a6fd37688e37dffedfcee7909dad831e9c9edb4"} Dec 04 11:58:13 crc kubenswrapper[4979]: I1204 11:58:13.910989 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" podStartSLOduration=2.315022466 podStartE2EDuration="7.910964389s" podCreationTimestamp="2025-12-04 11:58:06 +0000 UTC" firstStartedPulling="2025-12-04 11:58:07.541405956 +0000 UTC m=+911.815701760" lastFinishedPulling="2025-12-04 11:58:13.137283707 +0000 UTC m=+917.411643683" observedRunningTime="2025-12-04 11:58:13.905608304 +0000 UTC m=+918.179904138" watchObservedRunningTime="2025-12-04 11:58:13.910964389 +0000 UTC m=+918.185260223" Dec 04 11:58:14 crc kubenswrapper[4979]: I1204 11:58:14.895580 4979 generic.go:334] "Generic (PLEG): container finished" podID="40453e49-262b-47eb-ba13-a0de8531e17b" containerID="071004cec4bd3811dc2bf50cd449f03816136fc91f298cdbf56a485663f592e5" exitCode=0 Dec 04 11:58:14 crc kubenswrapper[4979]: I1204 11:58:14.895677 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4v6w8" event={"ID":"40453e49-262b-47eb-ba13-a0de8531e17b","Type":"ContainerDied","Data":"071004cec4bd3811dc2bf50cd449f03816136fc91f298cdbf56a485663f592e5"} Dec 04 11:58:15 crc kubenswrapper[4979]: I1204 11:58:15.906504 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4v6w8" event={"ID":"40453e49-262b-47eb-ba13-a0de8531e17b","Type":"ContainerStarted","Data":"7bd9dacd3fa29a23611e800bfd2d24aacda0ad370de98a641ef7e956c18fc77a"} Dec 04 11:58:21 crc kubenswrapper[4979]: I1204 11:58:21.724963 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:21 crc kubenswrapper[4979]: I1204 11:58:21.725514 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:21 crc kubenswrapper[4979]: I1204 11:58:21.768486 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:21 crc kubenswrapper[4979]: I1204 11:58:21.786268 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4v6w8" podStartSLOduration=9.060384223 podStartE2EDuration="10.786253723s" podCreationTimestamp="2025-12-04 11:58:11 +0000 UTC" firstStartedPulling="2025-12-04 11:58:13.890778749 +0000 UTC m=+918.165074573" lastFinishedPulling="2025-12-04 11:58:15.616648269 +0000 UTC m=+919.890944073" observedRunningTime="2025-12-04 11:58:15.927909057 +0000 UTC m=+920.202204871" watchObservedRunningTime="2025-12-04 11:58:21.786253723 +0000 UTC m=+926.060549527" Dec 04 11:58:21 crc kubenswrapper[4979]: I1204 11:58:21.976807 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:24 crc kubenswrapper[4979]: I1204 11:58:24.060642 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4v6w8"] Dec 04 11:58:24 crc kubenswrapper[4979]: I1204 11:58:24.061155 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4v6w8" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="registry-server" containerID="cri-o://7bd9dacd3fa29a23611e800bfd2d24aacda0ad370de98a641ef7e956c18fc77a" gracePeriod=2 Dec 04 11:58:25 crc kubenswrapper[4979]: I1204 11:58:25.960177 4979 generic.go:334] "Generic (PLEG): container finished" podID="40453e49-262b-47eb-ba13-a0de8531e17b" containerID="7bd9dacd3fa29a23611e800bfd2d24aacda0ad370de98a641ef7e956c18fc77a" exitCode=0 Dec 04 11:58:25 crc kubenswrapper[4979]: I1204 11:58:25.960384 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4v6w8" event={"ID":"40453e49-262b-47eb-ba13-a0de8531e17b","Type":"ContainerDied","Data":"7bd9dacd3fa29a23611e800bfd2d24aacda0ad370de98a641ef7e956c18fc77a"} Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.633797 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.765024 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52xmp\" (UniqueName: \"kubernetes.io/projected/40453e49-262b-47eb-ba13-a0de8531e17b-kube-api-access-52xmp\") pod \"40453e49-262b-47eb-ba13-a0de8531e17b\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.765147 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-catalog-content\") pod \"40453e49-262b-47eb-ba13-a0de8531e17b\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.765181 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-utilities\") pod \"40453e49-262b-47eb-ba13-a0de8531e17b\" (UID: \"40453e49-262b-47eb-ba13-a0de8531e17b\") " Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.766751 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-utilities" (OuterVolumeSpecName: "utilities") pod "40453e49-262b-47eb-ba13-a0de8531e17b" (UID: "40453e49-262b-47eb-ba13-a0de8531e17b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.776412 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40453e49-262b-47eb-ba13-a0de8531e17b-kube-api-access-52xmp" (OuterVolumeSpecName: "kube-api-access-52xmp") pod "40453e49-262b-47eb-ba13-a0de8531e17b" (UID: "40453e49-262b-47eb-ba13-a0de8531e17b"). InnerVolumeSpecName "kube-api-access-52xmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.843520 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "40453e49-262b-47eb-ba13-a0de8531e17b" (UID: "40453e49-262b-47eb-ba13-a0de8531e17b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.867375 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.867407 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/40453e49-262b-47eb-ba13-a0de8531e17b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.867419 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52xmp\" (UniqueName: \"kubernetes.io/projected/40453e49-262b-47eb-ba13-a0de8531e17b-kube-api-access-52xmp\") on node \"crc\" DevicePath \"\"" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.970368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4v6w8" event={"ID":"40453e49-262b-47eb-ba13-a0de8531e17b","Type":"ContainerDied","Data":"1e3a0e3e7e5af097bfd9a8cd1a6fd37688e37dffedfcee7909dad831e9c9edb4"} Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.970430 4979 scope.go:117] "RemoveContainer" containerID="7bd9dacd3fa29a23611e800bfd2d24aacda0ad370de98a641ef7e956c18fc77a" Dec 04 11:58:26 crc kubenswrapper[4979]: I1204 11:58:26.970463 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4v6w8" Dec 04 11:58:27 crc kubenswrapper[4979]: I1204 11:58:27.030442 4979 scope.go:117] "RemoveContainer" containerID="071004cec4bd3811dc2bf50cd449f03816136fc91f298cdbf56a485663f592e5" Dec 04 11:58:27 crc kubenswrapper[4979]: I1204 11:58:27.039386 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4v6w8"] Dec 04 11:58:27 crc kubenswrapper[4979]: I1204 11:58:27.094752 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4v6w8"] Dec 04 11:58:27 crc kubenswrapper[4979]: I1204 11:58:27.101922 4979 scope.go:117] "RemoveContainer" containerID="d2afe584aae1109d0aca0e2f22e9f5dcd0bb369c5b349fa801070f28ec54234b" Dec 04 11:58:27 crc kubenswrapper[4979]: I1204 11:58:27.133610 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-86498874f8-zlxh7" Dec 04 11:58:28 crc kubenswrapper[4979]: I1204 11:58:28.041742 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:58:28 crc kubenswrapper[4979]: I1204 11:58:28.042163 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:58:28 crc kubenswrapper[4979]: I1204 11:58:28.209040 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" path="/var/lib/kubelet/pods/40453e49-262b-47eb-ba13-a0de8531e17b/volumes" Dec 04 11:58:46 crc kubenswrapper[4979]: I1204 11:58:46.748720 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7d59d98bc-t8mjh" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.480832 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6"] Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.481142 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="registry-server" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.481174 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="registry-server" Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.481219 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="extract-content" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.481231 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="extract-content" Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.481252 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="extract-utilities" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.481265 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="extract-utilities" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.481489 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="40453e49-262b-47eb-ba13-a0de8531e17b" containerName="registry-server" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.482134 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.484129 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.484350 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-sbk2k"] Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.486276 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.488337 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.491137 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.498866 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-mvhcd" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.508584 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6"] Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.604963 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-msnnj"] Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.609227 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.610456 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-d4d2t"] Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.611273 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.612702 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.612814 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.612879 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.613004 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-25bmk" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.618334 4979 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.620693 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d4d2t"] Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642125 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-sockets\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642163 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-conf\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642184 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7760815-ac81-45eb-8a1a-0b6afe53f006-metrics-certs\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642209 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-metrics\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642238 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-reloader\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642252 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-startup\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642269 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wnsj\" (UniqueName: \"kubernetes.io/projected/b7760815-ac81-45eb-8a1a-0b6afe53f006-kube-api-access-9wnsj\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642289 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e84c35a-57f3-43e8-9d36-6e9750b85c9d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-cw6w6\" (UID: \"0e84c35a-57f3-43e8-9d36-6e9750b85c9d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.642331 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvzfz\" (UniqueName: \"kubernetes.io/projected/0e84c35a-57f3-43e8-9d36-6e9750b85c9d-kube-api-access-cvzfz\") pod \"frr-k8s-webhook-server-7fcb986d4-cw6w6\" (UID: \"0e84c35a-57f3-43e8-9d36-6e9750b85c9d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743774 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e84c35a-57f3-43e8-9d36-6e9750b85c9d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-cw6w6\" (UID: \"0e84c35a-57f3-43e8-9d36-6e9750b85c9d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743828 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvzfz\" (UniqueName: \"kubernetes.io/projected/0e84c35a-57f3-43e8-9d36-6e9750b85c9d-kube-api-access-cvzfz\") pod \"frr-k8s-webhook-server-7fcb986d4-cw6w6\" (UID: \"0e84c35a-57f3-43e8-9d36-6e9750b85c9d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743851 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xxx5\" (UniqueName: \"kubernetes.io/projected/f90cd359-ebe3-4c34-a2c0-16bf8805b876-kube-api-access-9xxx5\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743869 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-cert\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-sockets\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743925 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-conf\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.743947 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7760815-ac81-45eb-8a1a-0b6afe53f006-metrics-certs\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744512 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-sockets\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744567 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-conf\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744693 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-metrics\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744739 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-metrics-certs\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744763 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xpjc\" (UniqueName: \"kubernetes.io/projected/78803b74-9eb0-4721-91c5-39490f93248a-kube-api-access-9xpjc\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744802 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/78803b74-9eb0-4721-91c5-39490f93248a-metallb-excludel2\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744921 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744958 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-reloader\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.744986 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-startup\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.745002 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-metrics-certs\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.745026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wnsj\" (UniqueName: \"kubernetes.io/projected/b7760815-ac81-45eb-8a1a-0b6afe53f006-kube-api-access-9wnsj\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.745050 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-metrics\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.745345 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/b7760815-ac81-45eb-8a1a-0b6afe53f006-reloader\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.746030 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/b7760815-ac81-45eb-8a1a-0b6afe53f006-frr-startup\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.753987 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/b7760815-ac81-45eb-8a1a-0b6afe53f006-metrics-certs\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.761242 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvzfz\" (UniqueName: \"kubernetes.io/projected/0e84c35a-57f3-43e8-9d36-6e9750b85c9d-kube-api-access-cvzfz\") pod \"frr-k8s-webhook-server-7fcb986d4-cw6w6\" (UID: \"0e84c35a-57f3-43e8-9d36-6e9750b85c9d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.762149 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0e84c35a-57f3-43e8-9d36-6e9750b85c9d-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-cw6w6\" (UID: \"0e84c35a-57f3-43e8-9d36-6e9750b85c9d\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.763947 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wnsj\" (UniqueName: \"kubernetes.io/projected/b7760815-ac81-45eb-8a1a-0b6afe53f006-kube-api-access-9wnsj\") pod \"frr-k8s-sbk2k\" (UID: \"b7760815-ac81-45eb-8a1a-0b6afe53f006\") " pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.845976 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846236 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xxx5\" (UniqueName: \"kubernetes.io/projected/f90cd359-ebe3-4c34-a2c0-16bf8805b876-kube-api-access-9xxx5\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846278 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-cert\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846411 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-metrics-certs\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846441 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xpjc\" (UniqueName: \"kubernetes.io/projected/78803b74-9eb0-4721-91c5-39490f93248a-kube-api-access-9xpjc\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846476 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/78803b74-9eb0-4721-91c5-39490f93248a-metallb-excludel2\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846525 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.846558 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-metrics-certs\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.846700 4979 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.846767 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-metrics-certs podName:f90cd359-ebe3-4c34-a2c0-16bf8805b876 nodeName:}" failed. No retries permitted until 2025-12-04 11:58:48.346747805 +0000 UTC m=+952.621043609 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-metrics-certs") pod "controller-f8648f98b-d4d2t" (UID: "f90cd359-ebe3-4c34-a2c0-16bf8805b876") : secret "controller-certs-secret" not found Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.846773 4979 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 11:58:47 crc kubenswrapper[4979]: E1204 11:58:47.846836 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist podName:78803b74-9eb0-4721-91c5-39490f93248a nodeName:}" failed. No retries permitted until 2025-12-04 11:58:48.346816157 +0000 UTC m=+952.621111951 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist") pod "speaker-msnnj" (UID: "78803b74-9eb0-4721-91c5-39490f93248a") : secret "metallb-memberlist" not found Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.847346 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/78803b74-9eb0-4721-91c5-39490f93248a-metallb-excludel2\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.849496 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-cert\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.850705 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-metrics-certs\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.857601 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.872749 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xpjc\" (UniqueName: \"kubernetes.io/projected/78803b74-9eb0-4721-91c5-39490f93248a-kube-api-access-9xpjc\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:47 crc kubenswrapper[4979]: I1204 11:58:47.879863 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xxx5\" (UniqueName: \"kubernetes.io/projected/f90cd359-ebe3-4c34-a2c0-16bf8805b876-kube-api-access-9xxx5\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.094610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"9b8923413a8d8805d4ff2096ed56e039d7417535694d814d60a77cdbb7b0c009"} Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.291952 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6"] Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.357515 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.357815 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-metrics-certs\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:48 crc kubenswrapper[4979]: E1204 11:58:48.357672 4979 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 04 11:58:48 crc kubenswrapper[4979]: E1204 11:58:48.357904 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist podName:78803b74-9eb0-4721-91c5-39490f93248a nodeName:}" failed. No retries permitted until 2025-12-04 11:58:49.35788868 +0000 UTC m=+953.632184484 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist") pod "speaker-msnnj" (UID: "78803b74-9eb0-4721-91c5-39490f93248a") : secret "metallb-memberlist" not found Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.362131 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f90cd359-ebe3-4c34-a2c0-16bf8805b876-metrics-certs\") pod \"controller-f8648f98b-d4d2t\" (UID: \"f90cd359-ebe3-4c34-a2c0-16bf8805b876\") " pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.542996 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:48 crc kubenswrapper[4979]: I1204 11:58:48.796970 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-d4d2t"] Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.100031 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" event={"ID":"0e84c35a-57f3-43e8-9d36-6e9750b85c9d","Type":"ContainerStarted","Data":"ea438efd636891917fd489c12d5113d8758e4dbe287149ee77fdc8af23776acd"} Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.101752 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d4d2t" event={"ID":"f90cd359-ebe3-4c34-a2c0-16bf8805b876","Type":"ContainerStarted","Data":"1332116d99293a76109682e5a55d49e4af16858ea135d534fc2aba7fd74090bd"} Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.101777 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d4d2t" event={"ID":"f90cd359-ebe3-4c34-a2c0-16bf8805b876","Type":"ContainerStarted","Data":"49069a98eca97e8d39fd16a8d1f45a61050e0b9cc99d2f2c6581169a36f3a7ad"} Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.101788 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-d4d2t" event={"ID":"f90cd359-ebe3-4c34-a2c0-16bf8805b876","Type":"ContainerStarted","Data":"15386d57a83fb948d99ffba0f323068acd52b769ea2b6e5656d90b0f53a32ce0"} Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.101931 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.120344 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-d4d2t" podStartSLOduration=2.120290605 podStartE2EDuration="2.120290605s" podCreationTimestamp="2025-12-04 11:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:58:49.118868896 +0000 UTC m=+953.393164720" watchObservedRunningTime="2025-12-04 11:58:49.120290605 +0000 UTC m=+953.394586449" Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.374364 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.390006 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/78803b74-9eb0-4721-91c5-39490f93248a-memberlist\") pod \"speaker-msnnj\" (UID: \"78803b74-9eb0-4721-91c5-39490f93248a\") " pod="metallb-system/speaker-msnnj" Dec 04 11:58:49 crc kubenswrapper[4979]: I1204 11:58:49.430781 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-msnnj" Dec 04 11:58:50 crc kubenswrapper[4979]: I1204 11:58:50.108996 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-msnnj" event={"ID":"78803b74-9eb0-4721-91c5-39490f93248a","Type":"ContainerStarted","Data":"daee4d394e519acc54aa0c8b2c875d3e8074bdee954842e353268b6787d1e213"} Dec 04 11:58:50 crc kubenswrapper[4979]: I1204 11:58:50.109742 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-msnnj" event={"ID":"78803b74-9eb0-4721-91c5-39490f93248a","Type":"ContainerStarted","Data":"3d0e59087de59f3163c329a1329e83798ccb09a0a7982d75ef94410f81f7f4a8"} Dec 04 11:58:50 crc kubenswrapper[4979]: I1204 11:58:50.109758 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-msnnj" event={"ID":"78803b74-9eb0-4721-91c5-39490f93248a","Type":"ContainerStarted","Data":"0f0c68b062d1f892b548e1d99e9a457acb8141d4456097648bf3da1bd6a1c4e1"} Dec 04 11:58:50 crc kubenswrapper[4979]: I1204 11:58:50.109901 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-msnnj" Dec 04 11:58:50 crc kubenswrapper[4979]: I1204 11:58:50.128823 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-msnnj" podStartSLOduration=3.128806861 podStartE2EDuration="3.128806861s" podCreationTimestamp="2025-12-04 11:58:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:58:50.126257831 +0000 UTC m=+954.400553635" watchObservedRunningTime="2025-12-04 11:58:50.128806861 +0000 UTC m=+954.403102675" Dec 04 11:58:56 crc kubenswrapper[4979]: I1204 11:58:56.152611 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" event={"ID":"0e84c35a-57f3-43e8-9d36-6e9750b85c9d","Type":"ContainerStarted","Data":"91cbc3a2d86aab5f5672df34af01d0855f05702f6ba912fd24eae6a23019c5fe"} Dec 04 11:58:56 crc kubenswrapper[4979]: I1204 11:58:56.153088 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:58:56 crc kubenswrapper[4979]: I1204 11:58:56.153948 4979 generic.go:334] "Generic (PLEG): container finished" podID="b7760815-ac81-45eb-8a1a-0b6afe53f006" containerID="2487e4bb7628561b85cd6248c53a5f8dbebdea0eeba44ccfc816d16ec737ae52" exitCode=0 Dec 04 11:58:56 crc kubenswrapper[4979]: I1204 11:58:56.153975 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerDied","Data":"2487e4bb7628561b85cd6248c53a5f8dbebdea0eeba44ccfc816d16ec737ae52"} Dec 04 11:58:56 crc kubenswrapper[4979]: I1204 11:58:56.171034 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" podStartSLOduration=2.318800946 podStartE2EDuration="9.171020106s" podCreationTimestamp="2025-12-04 11:58:47 +0000 UTC" firstStartedPulling="2025-12-04 11:58:48.297353838 +0000 UTC m=+952.571649772" lastFinishedPulling="2025-12-04 11:58:55.149573128 +0000 UTC m=+959.423868932" observedRunningTime="2025-12-04 11:58:56.16932065 +0000 UTC m=+960.443616454" watchObservedRunningTime="2025-12-04 11:58:56.171020106 +0000 UTC m=+960.445315910" Dec 04 11:58:57 crc kubenswrapper[4979]: I1204 11:58:57.162432 4979 generic.go:334] "Generic (PLEG): container finished" podID="b7760815-ac81-45eb-8a1a-0b6afe53f006" containerID="5f2ba6bb9947b04f4579a90f02fb8ec3b94d14ba102bce6ec4af304772e42d06" exitCode=0 Dec 04 11:58:57 crc kubenswrapper[4979]: I1204 11:58:57.162481 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerDied","Data":"5f2ba6bb9947b04f4579a90f02fb8ec3b94d14ba102bce6ec4af304772e42d06"} Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.040733 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.041023 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.041078 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.041694 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.041758 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d" gracePeriod=600 Dec 04 11:58:58 crc kubenswrapper[4979]: E1204 11:58:58.152282 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9b48d97_5715_467e_b044_4812d6dfea63.slice/crio-8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.202387 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d" exitCode=0 Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.205275 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d"} Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.205340 4979 scope.go:117] "RemoveContainer" containerID="aa49c902e467cec7eb7ad9fa7d11c15836d2d5553c53624c2c9cd8f02a5491c5" Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.209411 4979 generic.go:334] "Generic (PLEG): container finished" podID="b7760815-ac81-45eb-8a1a-0b6afe53f006" containerID="a32483d512fa96c0555c36098702563eb627ad50a5e89fdb74e445b806124919" exitCode=0 Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.209448 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerDied","Data":"a32483d512fa96c0555c36098702563eb627ad50a5e89fdb74e445b806124919"} Dec 04 11:58:58 crc kubenswrapper[4979]: I1204 11:58:58.546711 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-d4d2t" Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.217963 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"bb5b5fb8f2704900b621034dde12cd64efd7f6ba3338a4dd8572ae9e51dbc1d8"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223233 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"625f2bd6df1010216981f200f449955103b56cfc8ff3b98c372df265600c874a"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223293 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"e2d5c5ef99a3340e9557fd3a532e049a69569543395c67ab141654394eb4bfec"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223334 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"db00de31040d321a080b1f8234000bffcc24abbc36e58dd629c35a7747f3fe91"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223344 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"3bfe5d6363bd46dc1ff9eb1088eee21b01705198e57bef95de7323efd11b5a8d"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223356 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"ffc092746944675b4fb3e582b38291ce1643e88f6b626c9930b89141afe9f757"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223364 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-sbk2k" event={"ID":"b7760815-ac81-45eb-8a1a-0b6afe53f006","Type":"ContainerStarted","Data":"be6b89bbf2f904589bca6ddf4daa70ea7dab7effd8e99a22a5a8391b5feb296c"} Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.223492 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.270906 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-sbk2k" podStartSLOduration=5.180199148 podStartE2EDuration="12.27088718s" podCreationTimestamp="2025-12-04 11:58:47 +0000 UTC" firstStartedPulling="2025-12-04 11:58:48.037127833 +0000 UTC m=+952.311423637" lastFinishedPulling="2025-12-04 11:58:55.127815855 +0000 UTC m=+959.402111669" observedRunningTime="2025-12-04 11:58:59.265984615 +0000 UTC m=+963.540280419" watchObservedRunningTime="2025-12-04 11:58:59.27088718 +0000 UTC m=+963.545182984" Dec 04 11:58:59 crc kubenswrapper[4979]: I1204 11:58:59.435765 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-msnnj" Dec 04 11:59:00 crc kubenswrapper[4979]: I1204 11:59:00.973556 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n"] Dec 04 11:59:00 crc kubenswrapper[4979]: I1204 11:59:00.975101 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:00 crc kubenswrapper[4979]: I1204 11:59:00.977093 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 11:59:00 crc kubenswrapper[4979]: I1204 11:59:00.992992 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n"] Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.169367 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.169424 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7hsv\" (UniqueName: \"kubernetes.io/projected/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-kube-api-access-x7hsv\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.169450 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.271025 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.271599 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7hsv\" (UniqueName: \"kubernetes.io/projected/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-kube-api-access-x7hsv\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.271838 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.271843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.272295 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.306737 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7hsv\" (UniqueName: \"kubernetes.io/projected/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-kube-api-access-x7hsv\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.602516 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:01 crc kubenswrapper[4979]: I1204 11:59:01.802884 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n"] Dec 04 11:59:02 crc kubenswrapper[4979]: I1204 11:59:02.242288 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" event={"ID":"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97","Type":"ContainerStarted","Data":"f325bd7e6dde820819e283fc6da890200cfc0ed2e598bc9c54bcdbb11854dce8"} Dec 04 11:59:02 crc kubenswrapper[4979]: I1204 11:59:02.858819 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:59:02 crc kubenswrapper[4979]: I1204 11:59:02.902911 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:59:03 crc kubenswrapper[4979]: I1204 11:59:03.252154 4979 generic.go:334] "Generic (PLEG): container finished" podID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerID="2ca5d917b1b2c88f9a36493c42b1661c51fbfbc601ada6e3d2d5804651601b42" exitCode=0 Dec 04 11:59:03 crc kubenswrapper[4979]: I1204 11:59:03.252206 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" event={"ID":"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97","Type":"ContainerDied","Data":"2ca5d917b1b2c88f9a36493c42b1661c51fbfbc601ada6e3d2d5804651601b42"} Dec 04 11:59:07 crc kubenswrapper[4979]: I1204 11:59:07.284402 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" event={"ID":"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97","Type":"ContainerDied","Data":"54823cd954f911fd06894e760ff31999653dde7a2f08864a3480043a12b37e92"} Dec 04 11:59:07 crc kubenswrapper[4979]: I1204 11:59:07.284222 4979 generic.go:334] "Generic (PLEG): container finished" podID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerID="54823cd954f911fd06894e760ff31999653dde7a2f08864a3480043a12b37e92" exitCode=0 Dec 04 11:59:07 crc kubenswrapper[4979]: I1204 11:59:07.875590 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-cw6w6" Dec 04 11:59:09 crc kubenswrapper[4979]: I1204 11:59:09.298009 4979 generic.go:334] "Generic (PLEG): container finished" podID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerID="b9f77cb663b341f59b65949fd0e34c39271e783f1c6a59ebc7f319e24b10a6d0" exitCode=0 Dec 04 11:59:09 crc kubenswrapper[4979]: I1204 11:59:09.298103 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" event={"ID":"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97","Type":"ContainerDied","Data":"b9f77cb663b341f59b65949fd0e34c39271e783f1c6a59ebc7f319e24b10a6d0"} Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.633645 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.819598 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-util\") pod \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.819733 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7hsv\" (UniqueName: \"kubernetes.io/projected/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-kube-api-access-x7hsv\") pod \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.819803 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-bundle\") pod \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\" (UID: \"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97\") " Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.826710 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-kube-api-access-x7hsv" (OuterVolumeSpecName: "kube-api-access-x7hsv") pod "4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" (UID: "4dc4b0b6-7e83-4e24-a772-a9607d3c8d97"). InnerVolumeSpecName "kube-api-access-x7hsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.833507 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-bundle" (OuterVolumeSpecName: "bundle") pod "4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" (UID: "4dc4b0b6-7e83-4e24-a772-a9607d3c8d97"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.845576 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-util" (OuterVolumeSpecName: "util") pod "4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" (UID: "4dc4b0b6-7e83-4e24-a772-a9607d3c8d97"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.920870 4979 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-util\") on node \"crc\" DevicePath \"\"" Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.920905 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7hsv\" (UniqueName: \"kubernetes.io/projected/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-kube-api-access-x7hsv\") on node \"crc\" DevicePath \"\"" Dec 04 11:59:10 crc kubenswrapper[4979]: I1204 11:59:10.920915 4979 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4dc4b0b6-7e83-4e24-a772-a9607d3c8d97-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 11:59:11 crc kubenswrapper[4979]: I1204 11:59:11.314048 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" event={"ID":"4dc4b0b6-7e83-4e24-a772-a9607d3c8d97","Type":"ContainerDied","Data":"f325bd7e6dde820819e283fc6da890200cfc0ed2e598bc9c54bcdbb11854dce8"} Dec 04 11:59:11 crc kubenswrapper[4979]: I1204 11:59:11.314095 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f325bd7e6dde820819e283fc6da890200cfc0ed2e598bc9c54bcdbb11854dce8" Dec 04 11:59:11 crc kubenswrapper[4979]: I1204 11:59:11.314105 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n" Dec 04 11:59:17 crc kubenswrapper[4979]: I1204 11:59:17.861774 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-sbk2k" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.692767 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj"] Dec 04 11:59:18 crc kubenswrapper[4979]: E1204 11:59:18.693345 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="util" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.693359 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="util" Dec 04 11:59:18 crc kubenswrapper[4979]: E1204 11:59:18.693371 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="pull" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.693378 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="pull" Dec 04 11:59:18 crc kubenswrapper[4979]: E1204 11:59:18.693390 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="extract" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.693398 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="extract" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.693520 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dc4b0b6-7e83-4e24-a772-a9607d3c8d97" containerName="extract" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.694031 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.697394 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.697616 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.697932 4979 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-mnh9d" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.711077 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj"] Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.747556 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/18de19a1-477a-421b-834e-f9875eef83c2-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n7gxj\" (UID: \"18de19a1-477a-421b-834e-f9875eef83c2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.747628 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42r92\" (UniqueName: \"kubernetes.io/projected/18de19a1-477a-421b-834e-f9875eef83c2-kube-api-access-42r92\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n7gxj\" (UID: \"18de19a1-477a-421b-834e-f9875eef83c2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.848696 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/18de19a1-477a-421b-834e-f9875eef83c2-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n7gxj\" (UID: \"18de19a1-477a-421b-834e-f9875eef83c2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.848752 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42r92\" (UniqueName: \"kubernetes.io/projected/18de19a1-477a-421b-834e-f9875eef83c2-kube-api-access-42r92\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n7gxj\" (UID: \"18de19a1-477a-421b-834e-f9875eef83c2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.849155 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/18de19a1-477a-421b-834e-f9875eef83c2-tmp\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n7gxj\" (UID: \"18de19a1-477a-421b-834e-f9875eef83c2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:18 crc kubenswrapper[4979]: I1204 11:59:18.874423 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42r92\" (UniqueName: \"kubernetes.io/projected/18de19a1-477a-421b-834e-f9875eef83c2-kube-api-access-42r92\") pod \"cert-manager-operator-controller-manager-64cf6dff88-n7gxj\" (UID: \"18de19a1-477a-421b-834e-f9875eef83c2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:19 crc kubenswrapper[4979]: I1204 11:59:19.008024 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" Dec 04 11:59:19 crc kubenswrapper[4979]: I1204 11:59:19.302381 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj"] Dec 04 11:59:19 crc kubenswrapper[4979]: W1204 11:59:19.317003 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18de19a1_477a_421b_834e_f9875eef83c2.slice/crio-7ff2b8ab94b60e7cd45142683e91b91af3f9e130beeb58571175848444db9603 WatchSource:0}: Error finding container 7ff2b8ab94b60e7cd45142683e91b91af3f9e130beeb58571175848444db9603: Status 404 returned error can't find the container with id 7ff2b8ab94b60e7cd45142683e91b91af3f9e130beeb58571175848444db9603 Dec 04 11:59:19 crc kubenswrapper[4979]: I1204 11:59:19.370208 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" event={"ID":"18de19a1-477a-421b-834e-f9875eef83c2","Type":"ContainerStarted","Data":"7ff2b8ab94b60e7cd45142683e91b91af3f9e130beeb58571175848444db9603"} Dec 04 11:59:26 crc kubenswrapper[4979]: I1204 11:59:26.435083 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" event={"ID":"18de19a1-477a-421b-834e-f9875eef83c2","Type":"ContainerStarted","Data":"89b7d1737667699733206995b3bd6e42dfdb29ab409a940134a4782ccea49e2e"} Dec 04 11:59:26 crc kubenswrapper[4979]: I1204 11:59:26.456456 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64cf6dff88-n7gxj" podStartSLOduration=1.9627483749999999 podStartE2EDuration="8.456439732s" podCreationTimestamp="2025-12-04 11:59:18 +0000 UTC" firstStartedPulling="2025-12-04 11:59:19.31964384 +0000 UTC m=+983.593939644" lastFinishedPulling="2025-12-04 11:59:25.813335197 +0000 UTC m=+990.087631001" observedRunningTime="2025-12-04 11:59:26.454705024 +0000 UTC m=+990.729000848" watchObservedRunningTime="2025-12-04 11:59:26.456439732 +0000 UTC m=+990.730735536" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.674686 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-svrq5"] Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.676900 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.680781 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.680855 4979 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-gwmnx" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.681204 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.690651 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-svrq5"] Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.748232 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af5b7f91-c64a-4f1b-8a48-3863730942c3-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-svrq5\" (UID: \"af5b7f91-c64a-4f1b-8a48-3863730942c3\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.748360 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxqkv\" (UniqueName: \"kubernetes.io/projected/af5b7f91-c64a-4f1b-8a48-3863730942c3-kube-api-access-bxqkv\") pod \"cert-manager-cainjector-855d9ccff4-svrq5\" (UID: \"af5b7f91-c64a-4f1b-8a48-3863730942c3\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.850228 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af5b7f91-c64a-4f1b-8a48-3863730942c3-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-svrq5\" (UID: \"af5b7f91-c64a-4f1b-8a48-3863730942c3\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.850335 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxqkv\" (UniqueName: \"kubernetes.io/projected/af5b7f91-c64a-4f1b-8a48-3863730942c3-kube-api-access-bxqkv\") pod \"cert-manager-cainjector-855d9ccff4-svrq5\" (UID: \"af5b7f91-c64a-4f1b-8a48-3863730942c3\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.874424 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af5b7f91-c64a-4f1b-8a48-3863730942c3-bound-sa-token\") pod \"cert-manager-cainjector-855d9ccff4-svrq5\" (UID: \"af5b7f91-c64a-4f1b-8a48-3863730942c3\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:32 crc kubenswrapper[4979]: I1204 11:59:32.874427 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxqkv\" (UniqueName: \"kubernetes.io/projected/af5b7f91-c64a-4f1b-8a48-3863730942c3-kube-api-access-bxqkv\") pod \"cert-manager-cainjector-855d9ccff4-svrq5\" (UID: \"af5b7f91-c64a-4f1b-8a48-3863730942c3\") " pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:33 crc kubenswrapper[4979]: I1204 11:59:33.001501 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" Dec 04 11:59:33 crc kubenswrapper[4979]: I1204 11:59:33.206666 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-855d9ccff4-svrq5"] Dec 04 11:59:33 crc kubenswrapper[4979]: I1204 11:59:33.474236 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" event={"ID":"af5b7f91-c64a-4f1b-8a48-3863730942c3","Type":"ContainerStarted","Data":"b260602cb5be930066129ee1d33f2023a40ba7397c7fc144d02e73ed1f405b62"} Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.288788 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-kqcwm"] Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.289478 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.291557 4979 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-72j9t" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.296516 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-kqcwm"] Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.367581 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b762418e-2e6a-4678-ad5f-7acac1219d5e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-kqcwm\" (UID: \"b762418e-2e6a-4678-ad5f-7acac1219d5e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.367653 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52bxb\" (UniqueName: \"kubernetes.io/projected/b762418e-2e6a-4678-ad5f-7acac1219d5e-kube-api-access-52bxb\") pod \"cert-manager-webhook-f4fb5df64-kqcwm\" (UID: \"b762418e-2e6a-4678-ad5f-7acac1219d5e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.468661 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b762418e-2e6a-4678-ad5f-7acac1219d5e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-kqcwm\" (UID: \"b762418e-2e6a-4678-ad5f-7acac1219d5e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.468728 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52bxb\" (UniqueName: \"kubernetes.io/projected/b762418e-2e6a-4678-ad5f-7acac1219d5e-kube-api-access-52bxb\") pod \"cert-manager-webhook-f4fb5df64-kqcwm\" (UID: \"b762418e-2e6a-4678-ad5f-7acac1219d5e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.492251 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52bxb\" (UniqueName: \"kubernetes.io/projected/b762418e-2e6a-4678-ad5f-7acac1219d5e-kube-api-access-52bxb\") pod \"cert-manager-webhook-f4fb5df64-kqcwm\" (UID: \"b762418e-2e6a-4678-ad5f-7acac1219d5e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.497454 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b762418e-2e6a-4678-ad5f-7acac1219d5e-bound-sa-token\") pod \"cert-manager-webhook-f4fb5df64-kqcwm\" (UID: \"b762418e-2e6a-4678-ad5f-7acac1219d5e\") " pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:34 crc kubenswrapper[4979]: I1204 11:59:34.623627 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:35 crc kubenswrapper[4979]: I1204 11:59:35.095515 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-f4fb5df64-kqcwm"] Dec 04 11:59:35 crc kubenswrapper[4979]: I1204 11:59:35.488898 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" event={"ID":"b762418e-2e6a-4678-ad5f-7acac1219d5e","Type":"ContainerStarted","Data":"d966b00b3709033635ca8317a08c1a869ad33ebaf6b83072ab0e3d10020c26fa"} Dec 04 11:59:41 crc kubenswrapper[4979]: I1204 11:59:41.535285 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" event={"ID":"af5b7f91-c64a-4f1b-8a48-3863730942c3","Type":"ContainerStarted","Data":"fca964ade22995773e05e4a3b84a7d38b0e17a7ea663d0b18be593faf9bbb3da"} Dec 04 11:59:41 crc kubenswrapper[4979]: I1204 11:59:41.538634 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" event={"ID":"b762418e-2e6a-4678-ad5f-7acac1219d5e","Type":"ContainerStarted","Data":"974caabfbc56961c8ecde6b5239601e071869c85089c3cef6eb50bf1c8a88f90"} Dec 04 11:59:41 crc kubenswrapper[4979]: I1204 11:59:41.538841 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:41 crc kubenswrapper[4979]: I1204 11:59:41.548854 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-855d9ccff4-svrq5" podStartSLOduration=1.97473188 podStartE2EDuration="9.548833336s" podCreationTimestamp="2025-12-04 11:59:32 +0000 UTC" firstStartedPulling="2025-12-04 11:59:33.217059545 +0000 UTC m=+997.491355349" lastFinishedPulling="2025-12-04 11:59:40.791160961 +0000 UTC m=+1005.065456805" observedRunningTime="2025-12-04 11:59:41.546857012 +0000 UTC m=+1005.821152816" watchObservedRunningTime="2025-12-04 11:59:41.548833336 +0000 UTC m=+1005.823129140" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.604507 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" podStartSLOduration=7.921218491 podStartE2EDuration="13.604491437s" podCreationTimestamp="2025-12-04 11:59:34 +0000 UTC" firstStartedPulling="2025-12-04 11:59:35.109277353 +0000 UTC m=+999.383573157" lastFinishedPulling="2025-12-04 11:59:40.792550299 +0000 UTC m=+1005.066846103" observedRunningTime="2025-12-04 11:59:41.566343606 +0000 UTC m=+1005.840639420" watchObservedRunningTime="2025-12-04 11:59:47.604491437 +0000 UTC m=+1011.878787241" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.609263 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4p5pm"] Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.610034 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.612379 4979 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-wp5nj" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.617426 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4p5pm"] Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.686861 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znq9f\" (UniqueName: \"kubernetes.io/projected/ff130626-7c18-4528-869b-0191bc75cca3-kube-api-access-znq9f\") pod \"cert-manager-86cb77c54b-4p5pm\" (UID: \"ff130626-7c18-4528-869b-0191bc75cca3\") " pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.686908 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff130626-7c18-4528-869b-0191bc75cca3-bound-sa-token\") pod \"cert-manager-86cb77c54b-4p5pm\" (UID: \"ff130626-7c18-4528-869b-0191bc75cca3\") " pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.787801 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff130626-7c18-4528-869b-0191bc75cca3-bound-sa-token\") pod \"cert-manager-86cb77c54b-4p5pm\" (UID: \"ff130626-7c18-4528-869b-0191bc75cca3\") " pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.787909 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znq9f\" (UniqueName: \"kubernetes.io/projected/ff130626-7c18-4528-869b-0191bc75cca3-kube-api-access-znq9f\") pod \"cert-manager-86cb77c54b-4p5pm\" (UID: \"ff130626-7c18-4528-869b-0191bc75cca3\") " pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.806785 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znq9f\" (UniqueName: \"kubernetes.io/projected/ff130626-7c18-4528-869b-0191bc75cca3-kube-api-access-znq9f\") pod \"cert-manager-86cb77c54b-4p5pm\" (UID: \"ff130626-7c18-4528-869b-0191bc75cca3\") " pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.806966 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ff130626-7c18-4528-869b-0191bc75cca3-bound-sa-token\") pod \"cert-manager-86cb77c54b-4p5pm\" (UID: \"ff130626-7c18-4528-869b-0191bc75cca3\") " pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:47 crc kubenswrapper[4979]: I1204 11:59:47.933050 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-86cb77c54b-4p5pm" Dec 04 11:59:48 crc kubenswrapper[4979]: I1204 11:59:48.157256 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-86cb77c54b-4p5pm"] Dec 04 11:59:48 crc kubenswrapper[4979]: I1204 11:59:48.578060 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-4p5pm" event={"ID":"ff130626-7c18-4528-869b-0191bc75cca3","Type":"ContainerStarted","Data":"5555ce54d78a07eccfbbeaa34d789fbf4c1c73487bd6df859035350d71f88ea8"} Dec 04 11:59:48 crc kubenswrapper[4979]: I1204 11:59:48.578242 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-86cb77c54b-4p5pm" event={"ID":"ff130626-7c18-4528-869b-0191bc75cca3","Type":"ContainerStarted","Data":"45d8e39ea40a80071d89dd540a87eaac850269ee9e9ba696e43a39b2991353b7"} Dec 04 11:59:48 crc kubenswrapper[4979]: I1204 11:59:48.597287 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-86cb77c54b-4p5pm" podStartSLOduration=1.597255295 podStartE2EDuration="1.597255295s" podCreationTimestamp="2025-12-04 11:59:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 11:59:48.596742191 +0000 UTC m=+1012.871038035" watchObservedRunningTime="2025-12-04 11:59:48.597255295 +0000 UTC m=+1012.871551139" Dec 04 11:59:49 crc kubenswrapper[4979]: I1204 11:59:49.626984 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-f4fb5df64-kqcwm" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.706033 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-lj9fp"] Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.707317 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.710381 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-2j5cc" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.711057 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.723848 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.729730 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lj9fp"] Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.876101 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rxl9\" (UniqueName: \"kubernetes.io/projected/56060e39-13b1-4058-a329-5b29c59aae9e-kube-api-access-5rxl9\") pod \"openstack-operator-index-lj9fp\" (UID: \"56060e39-13b1-4058-a329-5b29c59aae9e\") " pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.977500 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rxl9\" (UniqueName: \"kubernetes.io/projected/56060e39-13b1-4058-a329-5b29c59aae9e-kube-api-access-5rxl9\") pod \"openstack-operator-index-lj9fp\" (UID: \"56060e39-13b1-4058-a329-5b29c59aae9e\") " pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:52 crc kubenswrapper[4979]: I1204 11:59:52.996915 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rxl9\" (UniqueName: \"kubernetes.io/projected/56060e39-13b1-4058-a329-5b29c59aae9e-kube-api-access-5rxl9\") pod \"openstack-operator-index-lj9fp\" (UID: \"56060e39-13b1-4058-a329-5b29c59aae9e\") " pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:53 crc kubenswrapper[4979]: I1204 11:59:53.027084 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:53 crc kubenswrapper[4979]: I1204 11:59:53.427157 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-lj9fp"] Dec 04 11:59:53 crc kubenswrapper[4979]: W1204 11:59:53.431644 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod56060e39_13b1_4058_a329_5b29c59aae9e.slice/crio-124cde10539331a77c8c9812ef35a9604f0c3e465a96997025b86f2269a258ce WatchSource:0}: Error finding container 124cde10539331a77c8c9812ef35a9604f0c3e465a96997025b86f2269a258ce: Status 404 returned error can't find the container with id 124cde10539331a77c8c9812ef35a9604f0c3e465a96997025b86f2269a258ce Dec 04 11:59:53 crc kubenswrapper[4979]: I1204 11:59:53.608853 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lj9fp" event={"ID":"56060e39-13b1-4058-a329-5b29c59aae9e","Type":"ContainerStarted","Data":"124cde10539331a77c8c9812ef35a9604f0c3e465a96997025b86f2269a258ce"} Dec 04 11:59:55 crc kubenswrapper[4979]: I1204 11:59:55.488002 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lj9fp"] Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.094931 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xngxq"] Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.095940 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.106993 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xngxq"] Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.224341 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9mm4\" (UniqueName: \"kubernetes.io/projected/26a7a95d-a941-480c-b4a8-d2e892c26a45-kube-api-access-w9mm4\") pod \"openstack-operator-index-xngxq\" (UID: \"26a7a95d-a941-480c-b4a8-d2e892c26a45\") " pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.326025 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9mm4\" (UniqueName: \"kubernetes.io/projected/26a7a95d-a941-480c-b4a8-d2e892c26a45-kube-api-access-w9mm4\") pod \"openstack-operator-index-xngxq\" (UID: \"26a7a95d-a941-480c-b4a8-d2e892c26a45\") " pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.343829 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9mm4\" (UniqueName: \"kubernetes.io/projected/26a7a95d-a941-480c-b4a8-d2e892c26a45-kube-api-access-w9mm4\") pod \"openstack-operator-index-xngxq\" (UID: \"26a7a95d-a941-480c-b4a8-d2e892c26a45\") " pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.418147 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.631551 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lj9fp" event={"ID":"56060e39-13b1-4058-a329-5b29c59aae9e","Type":"ContainerStarted","Data":"41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514"} Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.631717 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-lj9fp" podUID="56060e39-13b1-4058-a329-5b29c59aae9e" containerName="registry-server" containerID="cri-o://41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514" gracePeriod=2 Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.653023 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-lj9fp" podStartSLOduration=2.219825689 podStartE2EDuration="4.652999842s" podCreationTimestamp="2025-12-04 11:59:52 +0000 UTC" firstStartedPulling="2025-12-04 11:59:53.433611312 +0000 UTC m=+1017.707907136" lastFinishedPulling="2025-12-04 11:59:55.866785485 +0000 UTC m=+1020.141081289" observedRunningTime="2025-12-04 11:59:56.651268345 +0000 UTC m=+1020.925564149" watchObservedRunningTime="2025-12-04 11:59:56.652999842 +0000 UTC m=+1020.927295646" Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.699729 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xngxq"] Dec 04 11:59:56 crc kubenswrapper[4979]: I1204 11:59:56.985326 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.134963 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rxl9\" (UniqueName: \"kubernetes.io/projected/56060e39-13b1-4058-a329-5b29c59aae9e-kube-api-access-5rxl9\") pod \"56060e39-13b1-4058-a329-5b29c59aae9e\" (UID: \"56060e39-13b1-4058-a329-5b29c59aae9e\") " Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.143577 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56060e39-13b1-4058-a329-5b29c59aae9e-kube-api-access-5rxl9" (OuterVolumeSpecName: "kube-api-access-5rxl9") pod "56060e39-13b1-4058-a329-5b29c59aae9e" (UID: "56060e39-13b1-4058-a329-5b29c59aae9e"). InnerVolumeSpecName "kube-api-access-5rxl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.236699 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rxl9\" (UniqueName: \"kubernetes.io/projected/56060e39-13b1-4058-a329-5b29c59aae9e-kube-api-access-5rxl9\") on node \"crc\" DevicePath \"\"" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.639766 4979 generic.go:334] "Generic (PLEG): container finished" podID="56060e39-13b1-4058-a329-5b29c59aae9e" containerID="41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514" exitCode=0 Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.639842 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lj9fp" event={"ID":"56060e39-13b1-4058-a329-5b29c59aae9e","Type":"ContainerDied","Data":"41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514"} Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.639871 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-lj9fp" event={"ID":"56060e39-13b1-4058-a329-5b29c59aae9e","Type":"ContainerDied","Data":"124cde10539331a77c8c9812ef35a9604f0c3e465a96997025b86f2269a258ce"} Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.639875 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-lj9fp" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.639890 4979 scope.go:117] "RemoveContainer" containerID="41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.642663 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xngxq" event={"ID":"26a7a95d-a941-480c-b4a8-d2e892c26a45","Type":"ContainerStarted","Data":"8b1c1096415d3adb1eaed3001a566d054c6d04758cd84db70421fd326a2a3222"} Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.642754 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xngxq" event={"ID":"26a7a95d-a941-480c-b4a8-d2e892c26a45","Type":"ContainerStarted","Data":"2507a0aa164233f9b026ce7a4c9cb70d8cae649bf61541710c33ae92cefc1112"} Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.665493 4979 scope.go:117] "RemoveContainer" containerID="41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514" Dec 04 11:59:57 crc kubenswrapper[4979]: E1204 11:59:57.666629 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514\": container with ID starting with 41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514 not found: ID does not exist" containerID="41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.666727 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514"} err="failed to get container status \"41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514\": rpc error: code = NotFound desc = could not find container \"41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514\": container with ID starting with 41d7c323ec3f5755b847661f677b572d28a35c33d4196d36ce858ff9f35b5514 not found: ID does not exist" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.674706 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xngxq" podStartSLOduration=1.6218248339999999 podStartE2EDuration="1.674687383s" podCreationTimestamp="2025-12-04 11:59:56 +0000 UTC" firstStartedPulling="2025-12-04 11:59:56.770830081 +0000 UTC m=+1021.045125885" lastFinishedPulling="2025-12-04 11:59:56.82369264 +0000 UTC m=+1021.097988434" observedRunningTime="2025-12-04 11:59:57.671829755 +0000 UTC m=+1021.946125559" watchObservedRunningTime="2025-12-04 11:59:57.674687383 +0000 UTC m=+1021.948983187" Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.687285 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-lj9fp"] Dec 04 11:59:57 crc kubenswrapper[4979]: I1204 11:59:57.691073 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-lj9fp"] Dec 04 11:59:58 crc kubenswrapper[4979]: I1204 11:59:58.206924 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56060e39-13b1-4058-a329-5b29c59aae9e" path="/var/lib/kubelet/pods/56060e39-13b1-4058-a329-5b29c59aae9e/volumes" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.148588 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p"] Dec 04 12:00:00 crc kubenswrapper[4979]: E1204 12:00:00.149213 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56060e39-13b1-4058-a329-5b29c59aae9e" containerName="registry-server" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.149229 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="56060e39-13b1-4058-a329-5b29c59aae9e" containerName="registry-server" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.149382 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="56060e39-13b1-4058-a329-5b29c59aae9e" containerName="registry-server" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.149854 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.151903 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.152621 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.157836 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p"] Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.280359 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a94915fc-616b-4eb7-9133-19411e6ab093-secret-volume\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.280405 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgzh6\" (UniqueName: \"kubernetes.io/projected/a94915fc-616b-4eb7-9133-19411e6ab093-kube-api-access-jgzh6\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.280433 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a94915fc-616b-4eb7-9133-19411e6ab093-config-volume\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.381858 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a94915fc-616b-4eb7-9133-19411e6ab093-secret-volume\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.381923 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgzh6\" (UniqueName: \"kubernetes.io/projected/a94915fc-616b-4eb7-9133-19411e6ab093-kube-api-access-jgzh6\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.381961 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a94915fc-616b-4eb7-9133-19411e6ab093-config-volume\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.386913 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a94915fc-616b-4eb7-9133-19411e6ab093-secret-volume\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.394413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a94915fc-616b-4eb7-9133-19411e6ab093-config-volume\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.402956 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgzh6\" (UniqueName: \"kubernetes.io/projected/a94915fc-616b-4eb7-9133-19411e6ab093-kube-api-access-jgzh6\") pod \"collect-profiles-29414160-q7t8p\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.477528 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:00 crc kubenswrapper[4979]: I1204 12:00:00.874752 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p"] Dec 04 12:00:00 crc kubenswrapper[4979]: W1204 12:00:00.880482 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda94915fc_616b_4eb7_9133_19411e6ab093.slice/crio-7550bfe7874f1d2898d02a71cf3fd021ca82512bfa111d4ac45606e7617388de WatchSource:0}: Error finding container 7550bfe7874f1d2898d02a71cf3fd021ca82512bfa111d4ac45606e7617388de: Status 404 returned error can't find the container with id 7550bfe7874f1d2898d02a71cf3fd021ca82512bfa111d4ac45606e7617388de Dec 04 12:00:01 crc kubenswrapper[4979]: I1204 12:00:01.671873 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" event={"ID":"a94915fc-616b-4eb7-9133-19411e6ab093","Type":"ContainerStarted","Data":"7550bfe7874f1d2898d02a71cf3fd021ca82512bfa111d4ac45606e7617388de"} Dec 04 12:00:02 crc kubenswrapper[4979]: I1204 12:00:02.680647 4979 generic.go:334] "Generic (PLEG): container finished" podID="a94915fc-616b-4eb7-9133-19411e6ab093" containerID="02caaddbcb1c5158efe7f727897e2e968d74aa9a753cf1f03267474ede0c3f89" exitCode=0 Dec 04 12:00:02 crc kubenswrapper[4979]: I1204 12:00:02.680751 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" event={"ID":"a94915fc-616b-4eb7-9133-19411e6ab093","Type":"ContainerDied","Data":"02caaddbcb1c5158efe7f727897e2e968d74aa9a753cf1f03267474ede0c3f89"} Dec 04 12:00:03 crc kubenswrapper[4979]: I1204 12:00:03.970865 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.131903 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a94915fc-616b-4eb7-9133-19411e6ab093-config-volume\") pod \"a94915fc-616b-4eb7-9133-19411e6ab093\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.132025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a94915fc-616b-4eb7-9133-19411e6ab093-secret-volume\") pod \"a94915fc-616b-4eb7-9133-19411e6ab093\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.132085 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgzh6\" (UniqueName: \"kubernetes.io/projected/a94915fc-616b-4eb7-9133-19411e6ab093-kube-api-access-jgzh6\") pod \"a94915fc-616b-4eb7-9133-19411e6ab093\" (UID: \"a94915fc-616b-4eb7-9133-19411e6ab093\") " Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.133242 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a94915fc-616b-4eb7-9133-19411e6ab093-config-volume" (OuterVolumeSpecName: "config-volume") pod "a94915fc-616b-4eb7-9133-19411e6ab093" (UID: "a94915fc-616b-4eb7-9133-19411e6ab093"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.137763 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a94915fc-616b-4eb7-9133-19411e6ab093-kube-api-access-jgzh6" (OuterVolumeSpecName: "kube-api-access-jgzh6") pod "a94915fc-616b-4eb7-9133-19411e6ab093" (UID: "a94915fc-616b-4eb7-9133-19411e6ab093"). InnerVolumeSpecName "kube-api-access-jgzh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.138529 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a94915fc-616b-4eb7-9133-19411e6ab093-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a94915fc-616b-4eb7-9133-19411e6ab093" (UID: "a94915fc-616b-4eb7-9133-19411e6ab093"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.233225 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a94915fc-616b-4eb7-9133-19411e6ab093-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.233255 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a94915fc-616b-4eb7-9133-19411e6ab093-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.233264 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgzh6\" (UniqueName: \"kubernetes.io/projected/a94915fc-616b-4eb7-9133-19411e6ab093-kube-api-access-jgzh6\") on node \"crc\" DevicePath \"\"" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.692561 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" event={"ID":"a94915fc-616b-4eb7-9133-19411e6ab093","Type":"ContainerDied","Data":"7550bfe7874f1d2898d02a71cf3fd021ca82512bfa111d4ac45606e7617388de"} Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.692598 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7550bfe7874f1d2898d02a71cf3fd021ca82512bfa111d4ac45606e7617388de" Dec 04 12:00:04 crc kubenswrapper[4979]: I1204 12:00:04.692596 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p" Dec 04 12:00:06 crc kubenswrapper[4979]: I1204 12:00:06.419510 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 12:00:06 crc kubenswrapper[4979]: I1204 12:00:06.420135 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 12:00:06 crc kubenswrapper[4979]: I1204 12:00:06.461877 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 12:00:06 crc kubenswrapper[4979]: I1204 12:00:06.725694 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xngxq" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.558886 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm"] Dec 04 12:00:13 crc kubenswrapper[4979]: E1204 12:00:13.559872 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a94915fc-616b-4eb7-9133-19411e6ab093" containerName="collect-profiles" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.559891 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a94915fc-616b-4eb7-9133-19411e6ab093" containerName="collect-profiles" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.560087 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a94915fc-616b-4eb7-9133-19411e6ab093" containerName="collect-profiles" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.561658 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.565343 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-5bzgg" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.578442 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm"] Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.664926 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-util\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.665437 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvzdh\" (UniqueName: \"kubernetes.io/projected/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-kube-api-access-zvzdh\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.665533 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-bundle\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.766476 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-util\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.766548 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvzdh\" (UniqueName: \"kubernetes.io/projected/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-kube-api-access-zvzdh\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.766582 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-bundle\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.766980 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-bundle\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.767156 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-util\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.787402 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvzdh\" (UniqueName: \"kubernetes.io/projected/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-kube-api-access-zvzdh\") pod \"7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:13 crc kubenswrapper[4979]: I1204 12:00:13.879374 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:14 crc kubenswrapper[4979]: I1204 12:00:14.277120 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm"] Dec 04 12:00:14 crc kubenswrapper[4979]: I1204 12:00:14.768735 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" event={"ID":"6c9ef7f7-e93d-4c20-b799-9531bc5627a8","Type":"ContainerStarted","Data":"1bb051f873d6f2e2015b205d864668d1e38a448b7e671fec6d87bbad4114031f"} Dec 04 12:00:16 crc kubenswrapper[4979]: I1204 12:00:16.784938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" event={"ID":"6c9ef7f7-e93d-4c20-b799-9531bc5627a8","Type":"ContainerStarted","Data":"3f4d31e7133a0626e0e5d18dbddfec002bbec5add8316bbfbb47c1441c387c0f"} Dec 04 12:00:17 crc kubenswrapper[4979]: I1204 12:00:17.793628 4979 generic.go:334] "Generic (PLEG): container finished" podID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerID="3f4d31e7133a0626e0e5d18dbddfec002bbec5add8316bbfbb47c1441c387c0f" exitCode=0 Dec 04 12:00:17 crc kubenswrapper[4979]: I1204 12:00:17.793700 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" event={"ID":"6c9ef7f7-e93d-4c20-b799-9531bc5627a8","Type":"ContainerDied","Data":"3f4d31e7133a0626e0e5d18dbddfec002bbec5add8316bbfbb47c1441c387c0f"} Dec 04 12:00:18 crc kubenswrapper[4979]: I1204 12:00:18.806375 4979 generic.go:334] "Generic (PLEG): container finished" podID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerID="a517cfac4d0ff43c7f1e54f9ec15ccb7b46f06fed39631660f1cca536d4a81b5" exitCode=0 Dec 04 12:00:18 crc kubenswrapper[4979]: I1204 12:00:18.806576 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" event={"ID":"6c9ef7f7-e93d-4c20-b799-9531bc5627a8","Type":"ContainerDied","Data":"a517cfac4d0ff43c7f1e54f9ec15ccb7b46f06fed39631660f1cca536d4a81b5"} Dec 04 12:00:19 crc kubenswrapper[4979]: I1204 12:00:19.817164 4979 generic.go:334] "Generic (PLEG): container finished" podID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerID="9d0954ee124281fe56f20fce18e1d07691faf6bf23f9afe03555645fc31eae23" exitCode=0 Dec 04 12:00:19 crc kubenswrapper[4979]: I1204 12:00:19.817265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" event={"ID":"6c9ef7f7-e93d-4c20-b799-9531bc5627a8","Type":"ContainerDied","Data":"9d0954ee124281fe56f20fce18e1d07691faf6bf23f9afe03555645fc31eae23"} Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.080543 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.197157 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-bundle\") pod \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.197255 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvzdh\" (UniqueName: \"kubernetes.io/projected/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-kube-api-access-zvzdh\") pod \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.197279 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-util\") pod \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\" (UID: \"6c9ef7f7-e93d-4c20-b799-9531bc5627a8\") " Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.198030 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-bundle" (OuterVolumeSpecName: "bundle") pod "6c9ef7f7-e93d-4c20-b799-9531bc5627a8" (UID: "6c9ef7f7-e93d-4c20-b799-9531bc5627a8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.202917 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-kube-api-access-zvzdh" (OuterVolumeSpecName: "kube-api-access-zvzdh") pod "6c9ef7f7-e93d-4c20-b799-9531bc5627a8" (UID: "6c9ef7f7-e93d-4c20-b799-9531bc5627a8"). InnerVolumeSpecName "kube-api-access-zvzdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.211972 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-util" (OuterVolumeSpecName: "util") pod "6c9ef7f7-e93d-4c20-b799-9531bc5627a8" (UID: "6c9ef7f7-e93d-4c20-b799-9531bc5627a8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.299027 4979 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.299073 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvzdh\" (UniqueName: \"kubernetes.io/projected/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-kube-api-access-zvzdh\") on node \"crc\" DevicePath \"\"" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.299086 4979 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6c9ef7f7-e93d-4c20-b799-9531bc5627a8-util\") on node \"crc\" DevicePath \"\"" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.835101 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" event={"ID":"6c9ef7f7-e93d-4c20-b799-9531bc5627a8","Type":"ContainerDied","Data":"1bb051f873d6f2e2015b205d864668d1e38a448b7e671fec6d87bbad4114031f"} Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.835522 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bb051f873d6f2e2015b205d864668d1e38a448b7e671fec6d87bbad4114031f" Dec 04 12:00:21 crc kubenswrapper[4979]: I1204 12:00:21.835172 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.759200 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck"] Dec 04 12:00:25 crc kubenswrapper[4979]: E1204 12:00:25.759808 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="extract" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.759824 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="extract" Dec 04 12:00:25 crc kubenswrapper[4979]: E1204 12:00:25.759845 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="pull" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.759852 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="pull" Dec 04 12:00:25 crc kubenswrapper[4979]: E1204 12:00:25.759873 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="util" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.759882 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="util" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.760028 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c9ef7f7-e93d-4c20-b799-9531bc5627a8" containerName="extract" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.760515 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.764574 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-nm9zc" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.782133 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck"] Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.864226 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb927\" (UniqueName: \"kubernetes.io/projected/c8d63092-cd3d-497c-87ea-4be3a31aff53-kube-api-access-bb927\") pod \"openstack-operator-controller-operator-56668d8dc4-9dsck\" (UID: \"c8d63092-cd3d-497c-87ea-4be3a31aff53\") " pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.965605 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb927\" (UniqueName: \"kubernetes.io/projected/c8d63092-cd3d-497c-87ea-4be3a31aff53-kube-api-access-bb927\") pod \"openstack-operator-controller-operator-56668d8dc4-9dsck\" (UID: \"c8d63092-cd3d-497c-87ea-4be3a31aff53\") " pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:25 crc kubenswrapper[4979]: I1204 12:00:25.988149 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb927\" (UniqueName: \"kubernetes.io/projected/c8d63092-cd3d-497c-87ea-4be3a31aff53-kube-api-access-bb927\") pod \"openstack-operator-controller-operator-56668d8dc4-9dsck\" (UID: \"c8d63092-cd3d-497c-87ea-4be3a31aff53\") " pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:26 crc kubenswrapper[4979]: I1204 12:00:26.079214 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:26 crc kubenswrapper[4979]: I1204 12:00:26.480951 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck"] Dec 04 12:00:26 crc kubenswrapper[4979]: I1204 12:00:26.870649 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" event={"ID":"c8d63092-cd3d-497c-87ea-4be3a31aff53","Type":"ContainerStarted","Data":"8fb64b6162f19840869c1e3b4b5e35f82b5336bf3eaaf2650ea8efe93adc6247"} Dec 04 12:00:31 crc kubenswrapper[4979]: I1204 12:00:31.906658 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" event={"ID":"c8d63092-cd3d-497c-87ea-4be3a31aff53","Type":"ContainerStarted","Data":"0c08947a30c2055c1e3d100d9556b9fe0895d84d100c4bc85990ce73e4f404d9"} Dec 04 12:00:31 crc kubenswrapper[4979]: I1204 12:00:31.907059 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:31 crc kubenswrapper[4979]: I1204 12:00:31.948114 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" podStartSLOduration=2.437001645 podStartE2EDuration="6.948084142s" podCreationTimestamp="2025-12-04 12:00:25 +0000 UTC" firstStartedPulling="2025-12-04 12:00:26.490522071 +0000 UTC m=+1050.764817875" lastFinishedPulling="2025-12-04 12:00:31.001604568 +0000 UTC m=+1055.275900372" observedRunningTime="2025-12-04 12:00:31.946120169 +0000 UTC m=+1056.220415983" watchObservedRunningTime="2025-12-04 12:00:31.948084142 +0000 UTC m=+1056.222379946" Dec 04 12:00:36 crc kubenswrapper[4979]: I1204 12:00:36.083466 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-56668d8dc4-9dsck" Dec 04 12:00:58 crc kubenswrapper[4979]: I1204 12:00:58.041341 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:00:58 crc kubenswrapper[4979]: I1204 12:00:58.042435 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.079450 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.081187 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.084914 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.085275 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-c7bl8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.085847 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.087325 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-s8n4q" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.095193 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.107721 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.108721 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.113674 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-5txb9" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.115651 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.133696 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.144994 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.146158 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.157539 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-65vxv" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.160386 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.161678 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.164051 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-w9v4b" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.176873 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.195860 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.196821 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.200233 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-4gb4m" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.209940 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwcfk\" (UniqueName: \"kubernetes.io/projected/cc4f8bbc-9399-4c17-a7a9-ec39a429813c-kube-api-access-nwcfk\") pod \"barbican-operator-controller-manager-7d9dfd778-5vbf8\" (UID: \"cc4f8bbc-9399-4c17-a7a9-ec39a429813c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.210010 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jwr2\" (UniqueName: \"kubernetes.io/projected/3369e6a8-2d6d-4c72-b96a-2cf650c4c257-kube-api-access-6jwr2\") pod \"designate-operator-controller-manager-78b4bc895b-kw2fk\" (UID: \"3369e6a8-2d6d-4c72-b96a-2cf650c4c257\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.210050 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9chzr\" (UniqueName: \"kubernetes.io/projected/aa22fecb-5237-48bd-9526-857132083b7f-kube-api-access-9chzr\") pod \"cinder-operator-controller-manager-859b6ccc6-lw7hk\" (UID: \"aa22fecb-5237-48bd-9526-857132083b7f\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.213425 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.216346 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.217289 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.223831 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-brwbb" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.224114 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.225277 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.246255 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.293422 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.305102 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.309091 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-ct2sc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311589 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311651 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xt74s\" (UniqueName: \"kubernetes.io/projected/01d00bed-f29a-4ada-98d1-18ac5935eb06-kube-api-access-xt74s\") pod \"heat-operator-controller-manager-5f64f6f8bb-hgslt\" (UID: \"01d00bed-f29a-4ada-98d1-18ac5935eb06\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311694 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwcfk\" (UniqueName: \"kubernetes.io/projected/cc4f8bbc-9399-4c17-a7a9-ec39a429813c-kube-api-access-nwcfk\") pod \"barbican-operator-controller-manager-7d9dfd778-5vbf8\" (UID: \"cc4f8bbc-9399-4c17-a7a9-ec39a429813c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311726 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km86v\" (UniqueName: \"kubernetes.io/projected/bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6-kube-api-access-km86v\") pod \"horizon-operator-controller-manager-68c6d99b8f-5rcqm\" (UID: \"bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311782 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jwr2\" (UniqueName: \"kubernetes.io/projected/3369e6a8-2d6d-4c72-b96a-2cf650c4c257-kube-api-access-6jwr2\") pod \"designate-operator-controller-manager-78b4bc895b-kw2fk\" (UID: \"3369e6a8-2d6d-4c72-b96a-2cf650c4c257\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311832 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npd65\" (UniqueName: \"kubernetes.io/projected/11b2439d-9796-4f42-94b4-f0903aa555d6-kube-api-access-npd65\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311852 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbtxh\" (UniqueName: \"kubernetes.io/projected/5190c637-8327-4a8b-b351-4a6e86f6d59a-kube-api-access-qbtxh\") pod \"glance-operator-controller-manager-77987cd8cd-mr2rj\" (UID: \"5190c637-8327-4a8b-b351-4a6e86f6d59a\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.311901 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9chzr\" (UniqueName: \"kubernetes.io/projected/aa22fecb-5237-48bd-9526-857132083b7f-kube-api-access-9chzr\") pod \"cinder-operator-controller-manager-859b6ccc6-lw7hk\" (UID: \"aa22fecb-5237-48bd-9526-857132083b7f\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.334364 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.335456 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.344028 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hgvtg" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.360994 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwcfk\" (UniqueName: \"kubernetes.io/projected/cc4f8bbc-9399-4c17-a7a9-ec39a429813c-kube-api-access-nwcfk\") pod \"barbican-operator-controller-manager-7d9dfd778-5vbf8\" (UID: \"cc4f8bbc-9399-4c17-a7a9-ec39a429813c\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.361879 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9chzr\" (UniqueName: \"kubernetes.io/projected/aa22fecb-5237-48bd-9526-857132083b7f-kube-api-access-9chzr\") pod \"cinder-operator-controller-manager-859b6ccc6-lw7hk\" (UID: \"aa22fecb-5237-48bd-9526-857132083b7f\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.363273 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.367994 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jwr2\" (UniqueName: \"kubernetes.io/projected/3369e6a8-2d6d-4c72-b96a-2cf650c4c257-kube-api-access-6jwr2\") pod \"designate-operator-controller-manager-78b4bc895b-kw2fk\" (UID: \"3369e6a8-2d6d-4c72-b96a-2cf650c4c257\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.394657 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.418060 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xt74s\" (UniqueName: \"kubernetes.io/projected/01d00bed-f29a-4ada-98d1-18ac5935eb06-kube-api-access-xt74s\") pod \"heat-operator-controller-manager-5f64f6f8bb-hgslt\" (UID: \"01d00bed-f29a-4ada-98d1-18ac5935eb06\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.418111 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km86v\" (UniqueName: \"kubernetes.io/projected/bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6-kube-api-access-km86v\") pod \"horizon-operator-controller-manager-68c6d99b8f-5rcqm\" (UID: \"bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.418153 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsbt6\" (UniqueName: \"kubernetes.io/projected/316f3ead-f9a1-44c5-84f9-5d8304a9f15b-kube-api-access-zsbt6\") pod \"ironic-operator-controller-manager-6c548fd776-7gb6r\" (UID: \"316f3ead-f9a1-44c5-84f9-5d8304a9f15b\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.418182 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npd65\" (UniqueName: \"kubernetes.io/projected/11b2439d-9796-4f42-94b4-f0903aa555d6-kube-api-access-npd65\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.418198 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbtxh\" (UniqueName: \"kubernetes.io/projected/5190c637-8327-4a8b-b351-4a6e86f6d59a-kube-api-access-qbtxh\") pod \"glance-operator-controller-manager-77987cd8cd-mr2rj\" (UID: \"5190c637-8327-4a8b-b351-4a6e86f6d59a\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.418241 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: E1204 12:01:14.418375 4979 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:14 crc kubenswrapper[4979]: E1204 12:01:14.418434 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert podName:11b2439d-9796-4f42-94b4-f0903aa555d6 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:14.918411759 +0000 UTC m=+1099.192707563 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert") pod "infra-operator-controller-manager-57548d458d-2pnbm" (UID: "11b2439d-9796-4f42-94b4-f0903aa555d6") : secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.419222 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.435811 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.440346 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.441578 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.442273 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npd65\" (UniqueName: \"kubernetes.io/projected/11b2439d-9796-4f42-94b4-f0903aa555d6-kube-api-access-npd65\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.445047 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xt74s\" (UniqueName: \"kubernetes.io/projected/01d00bed-f29a-4ada-98d1-18ac5935eb06-kube-api-access-xt74s\") pod \"heat-operator-controller-manager-5f64f6f8bb-hgslt\" (UID: \"01d00bed-f29a-4ada-98d1-18ac5935eb06\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.445221 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-mfxwq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.446184 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.454226 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbtxh\" (UniqueName: \"kubernetes.io/projected/5190c637-8327-4a8b-b351-4a6e86f6d59a-kube-api-access-qbtxh\") pod \"glance-operator-controller-manager-77987cd8cd-mr2rj\" (UID: \"5190c637-8327-4a8b-b351-4a6e86f6d59a\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.460134 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.460690 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km86v\" (UniqueName: \"kubernetes.io/projected/bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6-kube-api-access-km86v\") pod \"horizon-operator-controller-manager-68c6d99b8f-5rcqm\" (UID: \"bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.461952 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.464505 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-zs8sz" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.478658 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.492420 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.508052 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.519611 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsbt6\" (UniqueName: \"kubernetes.io/projected/316f3ead-f9a1-44c5-84f9-5d8304a9f15b-kube-api-access-zsbt6\") pod \"ironic-operator-controller-manager-6c548fd776-7gb6r\" (UID: \"316f3ead-f9a1-44c5-84f9-5d8304a9f15b\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.519660 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhwcg\" (UniqueName: \"kubernetes.io/projected/0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b-kube-api-access-zhwcg\") pod \"keystone-operator-controller-manager-7765d96ddf-q2kv7\" (UID: \"0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.521038 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.526175 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.527798 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.536056 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.537151 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-m2dqh" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.537925 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.539562 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsbt6\" (UniqueName: \"kubernetes.io/projected/316f3ead-f9a1-44c5-84f9-5d8304a9f15b-kube-api-access-zsbt6\") pod \"ironic-operator-controller-manager-6c548fd776-7gb6r\" (UID: \"316f3ead-f9a1-44c5-84f9-5d8304a9f15b\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.539694 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.541119 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.543417 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.543628 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-9gwgj" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.546149 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-blqv8" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.579530 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.621326 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8ppb\" (UniqueName: \"kubernetes.io/projected/465158ac-c7e8-4df8-97ee-231ba9ca3541-kube-api-access-f8ppb\") pod \"mariadb-operator-controller-manager-56bbcc9d85-tv9lc\" (UID: \"465158ac-c7e8-4df8-97ee-231ba9ca3541\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.621509 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc2gb\" (UniqueName: \"kubernetes.io/projected/d7efdf6b-7859-4750-b531-8bae5f2332fd-kube-api-access-fc2gb\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-xzkn7\" (UID: \"d7efdf6b-7859-4750-b531-8bae5f2332fd\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.621546 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52skd\" (UniqueName: \"kubernetes.io/projected/fe9dded8-cafd-469a-b65d-cfa21673e4b1-kube-api-access-52skd\") pod \"manila-operator-controller-manager-7c79b5df47-hwjs2\" (UID: \"fe9dded8-cafd-469a-b65d-cfa21673e4b1\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.621766 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhwcg\" (UniqueName: \"kubernetes.io/projected/0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b-kube-api-access-zhwcg\") pod \"keystone-operator-controller-manager-7765d96ddf-q2kv7\" (UID: \"0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.622112 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.621796 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87wlp\" (UniqueName: \"kubernetes.io/projected/4d6d4d26-72f0-4b63-979c-a533eb0dd53f-kube-api-access-87wlp\") pod \"nova-operator-controller-manager-697bc559fc-xwtfq\" (UID: \"4d6d4d26-72f0-4b63-979c-a533eb0dd53f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.630529 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.631407 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.646799 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhwcg\" (UniqueName: \"kubernetes.io/projected/0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b-kube-api-access-zhwcg\") pod \"keystone-operator-controller-manager-7765d96ddf-q2kv7\" (UID: \"0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.660392 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.661824 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.662091 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.662689 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.663684 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.667699 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.670408 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-pjfbl" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.673113 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-lksfm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.674150 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-scpzc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.674231 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.686588 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.691657 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.698037 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.699393 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.701558 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-phkrc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.702999 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.714246 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.715023 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.724816 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8ppb\" (UniqueName: \"kubernetes.io/projected/465158ac-c7e8-4df8-97ee-231ba9ca3541-kube-api-access-f8ppb\") pod \"mariadb-operator-controller-manager-56bbcc9d85-tv9lc\" (UID: \"465158ac-c7e8-4df8-97ee-231ba9ca3541\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.724856 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52skd\" (UniqueName: \"kubernetes.io/projected/fe9dded8-cafd-469a-b65d-cfa21673e4b1-kube-api-access-52skd\") pod \"manila-operator-controller-manager-7c79b5df47-hwjs2\" (UID: \"fe9dded8-cafd-469a-b65d-cfa21673e4b1\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.724879 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc2gb\" (UniqueName: \"kubernetes.io/projected/d7efdf6b-7859-4750-b531-8bae5f2332fd-kube-api-access-fc2gb\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-xzkn7\" (UID: \"d7efdf6b-7859-4750-b531-8bae5f2332fd\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.725125 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87wlp\" (UniqueName: \"kubernetes.io/projected/4d6d4d26-72f0-4b63-979c-a533eb0dd53f-kube-api-access-87wlp\") pod \"nova-operator-controller-manager-697bc559fc-xwtfq\" (UID: \"4d6d4d26-72f0-4b63-979c-a533eb0dd53f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.725174 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqkn4\" (UniqueName: \"kubernetes.io/projected/e729014a-4926-4219-9159-785c4f602117-kube-api-access-rqkn4\") pod \"octavia-operator-controller-manager-998648c74-4j4cc\" (UID: \"e729014a-4926-4219-9159-785c4f602117\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.736752 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.739036 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.745552 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-nthfq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.749835 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8ppb\" (UniqueName: \"kubernetes.io/projected/465158ac-c7e8-4df8-97ee-231ba9ca3541-kube-api-access-f8ppb\") pod \"mariadb-operator-controller-manager-56bbcc9d85-tv9lc\" (UID: \"465158ac-c7e8-4df8-97ee-231ba9ca3541\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.754786 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.755028 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87wlp\" (UniqueName: \"kubernetes.io/projected/4d6d4d26-72f0-4b63-979c-a533eb0dd53f-kube-api-access-87wlp\") pod \"nova-operator-controller-manager-697bc559fc-xwtfq\" (UID: \"4d6d4d26-72f0-4b63-979c-a533eb0dd53f\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.758612 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc2gb\" (UniqueName: \"kubernetes.io/projected/d7efdf6b-7859-4750-b531-8bae5f2332fd-kube-api-access-fc2gb\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-xzkn7\" (UID: \"d7efdf6b-7859-4750-b531-8bae5f2332fd\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.758631 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52skd\" (UniqueName: \"kubernetes.io/projected/fe9dded8-cafd-469a-b65d-cfa21673e4b1-kube-api-access-52skd\") pod \"manila-operator-controller-manager-7c79b5df47-hwjs2\" (UID: \"fe9dded8-cafd-469a-b65d-cfa21673e4b1\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.786997 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.788519 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.792732 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-w6997" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.795904 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.805781 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.824211 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.824845 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829366 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc6k8\" (UniqueName: \"kubernetes.io/projected/3583aaeb-1aeb-4afc-91d2-a6837635a371-kube-api-access-cc6k8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829506 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqkn4\" (UniqueName: \"kubernetes.io/projected/e729014a-4926-4219-9159-785c4f602117-kube-api-access-rqkn4\") pod \"octavia-operator-controller-manager-998648c74-4j4cc\" (UID: \"e729014a-4926-4219-9159-785c4f602117\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829668 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ww9vj\" (UniqueName: \"kubernetes.io/projected/91e3f7ac-b8d9-40d8-8a1c-94412c0a9488-kube-api-access-ww9vj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-55g6q\" (UID: \"91e3f7ac-b8d9-40d8-8a1c-94412c0a9488\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829738 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp7p2\" (UniqueName: \"kubernetes.io/projected/30066915-9744-4901-82c8-d19d06b1fb49-kube-api-access-lp7p2\") pod \"placement-operator-controller-manager-78f8948974-7jr2r\" (UID: \"30066915-9744-4901-82c8-d19d06b1fb49\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829813 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t46fh\" (UniqueName: \"kubernetes.io/projected/7c730311-2e6a-4815-87f4-f8099d5ced63-kube-api-access-t46fh\") pod \"ovn-operator-controller-manager-b6456fdb6-xgjxw\" (UID: \"7c730311-2e6a-4815-87f4-f8099d5ced63\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829859 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf4pk\" (UniqueName: \"kubernetes.io/projected/d8e79a30-90c1-409c-a12b-ad2d0e0d36c8-kube-api-access-wf4pk\") pod \"swift-operator-controller-manager-5f8c65bbfc-dhmqs\" (UID: \"d8e79a30-90c1-409c-a12b-ad2d0e0d36c8\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.829940 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.831682 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.835802 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-59wkq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.852196 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.858058 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.871772 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqkn4\" (UniqueName: \"kubernetes.io/projected/e729014a-4926-4219-9159-785c4f602117-kube-api-access-rqkn4\") pod \"octavia-operator-controller-manager-998648c74-4j4cc\" (UID: \"e729014a-4926-4219-9159-785c4f602117\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.882873 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.899254 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931043 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf4pk\" (UniqueName: \"kubernetes.io/projected/d8e79a30-90c1-409c-a12b-ad2d0e0d36c8-kube-api-access-wf4pk\") pod \"swift-operator-controller-manager-5f8c65bbfc-dhmqs\" (UID: \"d8e79a30-90c1-409c-a12b-ad2d0e0d36c8\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931087 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931131 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd5sb\" (UniqueName: \"kubernetes.io/projected/fa56245e-7ded-4797-9fd9-bc84274407e2-kube-api-access-sd5sb\") pod \"test-operator-controller-manager-5854674fcc-c9c8v\" (UID: \"fa56245e-7ded-4797-9fd9-bc84274407e2\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc6k8\" (UniqueName: \"kubernetes.io/projected/3583aaeb-1aeb-4afc-91d2-a6837635a371-kube-api-access-cc6k8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931193 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dc4rc\" (UniqueName: \"kubernetes.io/projected/4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04-kube-api-access-dc4rc\") pod \"watcher-operator-controller-manager-769dc69bc-ddkhk\" (UID: \"4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931214 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931236 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ww9vj\" (UniqueName: \"kubernetes.io/projected/91e3f7ac-b8d9-40d8-8a1c-94412c0a9488-kube-api-access-ww9vj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-55g6q\" (UID: \"91e3f7ac-b8d9-40d8-8a1c-94412c0a9488\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931264 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp7p2\" (UniqueName: \"kubernetes.io/projected/30066915-9744-4901-82c8-d19d06b1fb49-kube-api-access-lp7p2\") pod \"placement-operator-controller-manager-78f8948974-7jr2r\" (UID: \"30066915-9744-4901-82c8-d19d06b1fb49\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:14 crc kubenswrapper[4979]: E1204 12:01:14.931283 4979 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:14 crc kubenswrapper[4979]: E1204 12:01:14.931443 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert podName:3583aaeb-1aeb-4afc-91d2-a6837635a371 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:15.43142247 +0000 UTC m=+1099.705718334 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" (UID: "3583aaeb-1aeb-4afc-91d2-a6837635a371") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:14 crc kubenswrapper[4979]: E1204 12:01:14.931575 4979 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.931290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t46fh\" (UniqueName: \"kubernetes.io/projected/7c730311-2e6a-4815-87f4-f8099d5ced63-kube-api-access-t46fh\") pod \"ovn-operator-controller-manager-b6456fdb6-xgjxw\" (UID: \"7c730311-2e6a-4815-87f4-f8099d5ced63\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:14 crc kubenswrapper[4979]: E1204 12:01:14.931618 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert podName:11b2439d-9796-4f42-94b4-f0903aa555d6 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:15.931603174 +0000 UTC m=+1100.205898978 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert") pod "infra-operator-controller-manager-57548d458d-2pnbm" (UID: "11b2439d-9796-4f42-94b4-f0903aa555d6") : secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.961264 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t46fh\" (UniqueName: \"kubernetes.io/projected/7c730311-2e6a-4815-87f4-f8099d5ced63-kube-api-access-t46fh\") pod \"ovn-operator-controller-manager-b6456fdb6-xgjxw\" (UID: \"7c730311-2e6a-4815-87f4-f8099d5ced63\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.962739 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf4pk\" (UniqueName: \"kubernetes.io/projected/d8e79a30-90c1-409c-a12b-ad2d0e0d36c8-kube-api-access-wf4pk\") pod \"swift-operator-controller-manager-5f8c65bbfc-dhmqs\" (UID: \"d8e79a30-90c1-409c-a12b-ad2d0e0d36c8\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.962830 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc6k8\" (UniqueName: \"kubernetes.io/projected/3583aaeb-1aeb-4afc-91d2-a6837635a371-kube-api-access-cc6k8\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.962957 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp7p2\" (UniqueName: \"kubernetes.io/projected/30066915-9744-4901-82c8-d19d06b1fb49-kube-api-access-lp7p2\") pod \"placement-operator-controller-manager-78f8948974-7jr2r\" (UID: \"30066915-9744-4901-82c8-d19d06b1fb49\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.964498 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ww9vj\" (UniqueName: \"kubernetes.io/projected/91e3f7ac-b8d9-40d8-8a1c-94412c0a9488-kube-api-access-ww9vj\") pod \"telemetry-operator-controller-manager-76cc84c6bb-55g6q\" (UID: \"91e3f7ac-b8d9-40d8-8a1c-94412c0a9488\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.991387 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l"] Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.992450 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.996892 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-wcvzp" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.997089 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.997194 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 04 12:01:14 crc kubenswrapper[4979]: I1204 12:01:14.997719 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.035799 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd5sb\" (UniqueName: \"kubernetes.io/projected/fa56245e-7ded-4797-9fd9-bc84274407e2-kube-api-access-sd5sb\") pod \"test-operator-controller-manager-5854674fcc-c9c8v\" (UID: \"fa56245e-7ded-4797-9fd9-bc84274407e2\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.035892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dc4rc\" (UniqueName: \"kubernetes.io/projected/4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04-kube-api-access-dc4rc\") pod \"watcher-operator-controller-manager-769dc69bc-ddkhk\" (UID: \"4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.041356 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.043661 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.045524 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.050726 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-8rwrq" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.056352 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.073076 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd5sb\" (UniqueName: \"kubernetes.io/projected/fa56245e-7ded-4797-9fd9-bc84274407e2-kube-api-access-sd5sb\") pod \"test-operator-controller-manager-5854674fcc-c9c8v\" (UID: \"fa56245e-7ded-4797-9fd9-bc84274407e2\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.073849 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dc4rc\" (UniqueName: \"kubernetes.io/projected/4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04-kube-api-access-dc4rc\") pod \"watcher-operator-controller-manager-769dc69bc-ddkhk\" (UID: \"4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:15 crc kubenswrapper[4979]: W1204 12:01:15.100767 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcc4f8bbc_9399_4c17_a7a9_ec39a429813c.slice/crio-91faa9fbe373ee91146d6e308adbfb63533bc4c4454e9f98d5927d0b6102a2ec WatchSource:0}: Error finding container 91faa9fbe373ee91146d6e308adbfb63533bc4c4454e9f98d5927d0b6102a2ec: Status 404 returned error can't find the container with id 91faa9fbe373ee91146d6e308adbfb63533bc4c4454e9f98d5927d0b6102a2ec Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.137273 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf488\" (UniqueName: \"kubernetes.io/projected/e5a502b0-c2bf-43fb-87eb-8def7494a0b9-kube-api-access-rf488\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kk4p2\" (UID: \"e5a502b0-c2bf-43fb-87eb-8def7494a0b9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.137649 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.137673 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hxvn\" (UniqueName: \"kubernetes.io/projected/8ef350ec-4178-410b-9e8e-9981810dc148-kube-api-access-4hxvn\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.137718 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.167896 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.177159 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" event={"ID":"cc4f8bbc-9399-4c17-a7a9-ec39a429813c","Type":"ContainerStarted","Data":"91faa9fbe373ee91146d6e308adbfb63533bc4c4454e9f98d5927d0b6102a2ec"} Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.217214 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.230341 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.238951 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.239084 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf488\" (UniqueName: \"kubernetes.io/projected/e5a502b0-c2bf-43fb-87eb-8def7494a0b9-kube-api-access-rf488\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kk4p2\" (UID: \"e5a502b0-c2bf-43fb-87eb-8def7494a0b9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.239118 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.239150 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hxvn\" (UniqueName: \"kubernetes.io/projected/8ef350ec-4178-410b-9e8e-9981810dc148-kube-api-access-4hxvn\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.239681 4979 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.239731 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:15.739711961 +0000 UTC m=+1100.014007765 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "metrics-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.240103 4979 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.240134 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:15.740124463 +0000 UTC m=+1100.014420267 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.269901 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.269991 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf488\" (UniqueName: \"kubernetes.io/projected/e5a502b0-c2bf-43fb-87eb-8def7494a0b9-kube-api-access-rf488\") pod \"rabbitmq-cluster-operator-manager-668c99d594-kk4p2\" (UID: \"e5a502b0-c2bf-43fb-87eb-8def7494a0b9\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.281133 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.294104 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.322406 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.322935 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.328473 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hxvn\" (UniqueName: \"kubernetes.io/projected/8ef350ec-4178-410b-9e8e-9981810dc148-kube-api-access-4hxvn\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.404772 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.449833 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.450056 4979 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.450183 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert podName:3583aaeb-1aeb-4afc-91d2-a6837635a371 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:16.450141506 +0000 UTC m=+1100.724437480 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" (UID: "3583aaeb-1aeb-4afc-91d2-a6837635a371") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: W1204 12:01:15.452921 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3369e6a8_2d6d_4c72_b96a_2cf650c4c257.slice/crio-718beaef22e4a00ea831b5cfdc218ec976ab538026191347f2be970c13fc30cd WatchSource:0}: Error finding container 718beaef22e4a00ea831b5cfdc218ec976ab538026191347f2be970c13fc30cd: Status 404 returned error can't find the container with id 718beaef22e4a00ea831b5cfdc218ec976ab538026191347f2be970c13fc30cd Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.455331 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.552822 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.589205 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.629160 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.755887 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.755998 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.756206 4979 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.756265 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:16.756247968 +0000 UTC m=+1101.030543772 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "metrics-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.756650 4979 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.756684 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:16.75667413 +0000 UTC m=+1101.030969944 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: W1204 12:01:15.764686 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd63d5e5_4cb3_4749_8f86_88a32e3ba6a6.slice/crio-b01f952685e1b4e0742f2dada53f5d4f8c9c9bfc8390183b5522b912e565a212 WatchSource:0}: Error finding container b01f952685e1b4e0742f2dada53f5d4f8c9c9bfc8390183b5522b912e565a212: Status 404 returned error can't find the container with id b01f952685e1b4e0742f2dada53f5d4f8c9c9bfc8390183b5522b912e565a212 Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.840827 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.877336 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.886747 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r"] Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.959039 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.959225 4979 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: E1204 12:01:15.959278 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert podName:11b2439d-9796-4f42-94b4-f0903aa555d6 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:17.959263551 +0000 UTC m=+1102.233559345 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert") pod "infra-operator-controller-manager-57548d458d-2pnbm" (UID: "11b2439d-9796-4f42-94b4-f0903aa555d6") : secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:15 crc kubenswrapper[4979]: I1204 12:01:15.959908 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq"] Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.039887 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7"] Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.077346 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r"] Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.090711 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc"] Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.106851 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lp7p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7jr2r_openstack-operators(30066915-9744-4901-82c8-d19d06b1fb49): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.109887 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lp7p2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-7jr2r_openstack-operators(30066915-9744-4901-82c8-d19d06b1fb49): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.112530 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" podUID="30066915-9744-4901-82c8-d19d06b1fb49" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.154189 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw"] Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.190460 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q"] Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.214593 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ww9vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-55g6q_openstack-operators(91e3f7ac-b8d9-40d8-8a1c-94412c0a9488): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.217204 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ww9vj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-55g6q_openstack-operators(91e3f7ac-b8d9-40d8-8a1c-94412c0a9488): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.218817 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" podUID="91e3f7ac-b8d9-40d8-8a1c-94412c0a9488" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.255207 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" event={"ID":"465158ac-c7e8-4df8-97ee-231ba9ca3541","Type":"ContainerStarted","Data":"8a16c8c904f90c418b767a7413c18894a07b5943732c9e6abe780c7011d77015"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.255246 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" event={"ID":"7c730311-2e6a-4815-87f4-f8099d5ced63","Type":"ContainerStarted","Data":"c019d31bb9cbf4ef3209e4ba209216481ffc37abfc0ae8ef47954ff58e10098e"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.255258 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" event={"ID":"316f3ead-f9a1-44c5-84f9-5d8304a9f15b","Type":"ContainerStarted","Data":"b4efc5fb56982ade1d9ad31fdc2cc1e1a30adc6a97c344563f6b0c5b68858117"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.256406 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" event={"ID":"d7efdf6b-7859-4750-b531-8bae5f2332fd","Type":"ContainerStarted","Data":"9b948c03bb0b0bc5bcfdcc81060c955c087bb3635da03564bf2ac45123aa6fea"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.257813 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" event={"ID":"3369e6a8-2d6d-4c72-b96a-2cf650c4c257","Type":"ContainerStarted","Data":"718beaef22e4a00ea831b5cfdc218ec976ab538026191347f2be970c13fc30cd"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.258603 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" event={"ID":"30066915-9744-4901-82c8-d19d06b1fb49","Type":"ContainerStarted","Data":"ddad4cb36a2c6aa2fde5bf0dad61df6d82bad2603b35f3e3ed222aaada6c363f"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.262393 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" event={"ID":"01d00bed-f29a-4ada-98d1-18ac5935eb06","Type":"ContainerStarted","Data":"a65546259a9813094040ccfde516444651fef11fa53db62ef78c2a4cd6f0682e"} Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.262484 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" podUID="30066915-9744-4901-82c8-d19d06b1fb49" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.276411 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" event={"ID":"4d6d4d26-72f0-4b63-979c-a533eb0dd53f","Type":"ContainerStarted","Data":"0fc2c84927b3e94795a8cc856a7fa60cff7d7b4b019145b0dc66b971c188b9cf"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.284612 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" event={"ID":"5190c637-8327-4a8b-b351-4a6e86f6d59a","Type":"ContainerStarted","Data":"e0fa5d363ce9e754be2edd586306d3fe40114194354bfce5384839db2eb38a6a"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.288011 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" event={"ID":"bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6","Type":"ContainerStarted","Data":"b01f952685e1b4e0742f2dada53f5d4f8c9c9bfc8390183b5522b912e565a212"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.290972 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" event={"ID":"0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b","Type":"ContainerStarted","Data":"9aefb3b1cce40f9b9eab8b0d846e8ce704a0eb99eda478ad45ec42ee76c4634a"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.294269 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" event={"ID":"fe9dded8-cafd-469a-b65d-cfa21673e4b1","Type":"ContainerStarted","Data":"8b184136f16abab7e607db96c5dddb1fb8ca97594888fba4cb1d8dbc6adad555"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.295201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" event={"ID":"aa22fecb-5237-48bd-9526-857132083b7f","Type":"ContainerStarted","Data":"6a6e952ad0d75e8bd5d35bbc67ba11924d99d3c3fde5080c6107b3ac46176fc4"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.296432 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" event={"ID":"e729014a-4926-4219-9159-785c4f602117","Type":"ContainerStarted","Data":"58bd50eac359da435f4828bcb5f9cd9fd7837ef142cb154c85bc06a97ddb21f7"} Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.331318 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2"] Dec 04 12:01:16 crc kubenswrapper[4979]: W1204 12:01:16.336713 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode5a502b0_c2bf_43fb_87eb_8def7494a0b9.slice/crio-efe12dc35f79bf9fcec65c1e88bee1530a0ef1111ca29d3808d99c42cb4a47b2 WatchSource:0}: Error finding container efe12dc35f79bf9fcec65c1e88bee1530a0ef1111ca29d3808d99c42cb4a47b2: Status 404 returned error can't find the container with id efe12dc35f79bf9fcec65c1e88bee1530a0ef1111ca29d3808d99c42cb4a47b2 Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.375077 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs"] Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.388826 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wf4pk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-dhmqs_openstack-operators(d8e79a30-90c1-409c-a12b-ad2d0e0d36c8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.390942 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wf4pk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-dhmqs_openstack-operators(d8e79a30-90c1-409c-a12b-ad2d0e0d36c8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.392294 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" podUID="d8e79a30-90c1-409c-a12b-ad2d0e0d36c8" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.456719 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk"] Dec 04 12:01:16 crc kubenswrapper[4979]: W1204 12:01:16.465610 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c3301d5_6fdd_4e3c_a4ba_f2c11322fd04.slice/crio-a9cdc95bd12b4affa5977bc2df1cd089d7c82f225e74e0a87f1e3fce43707bb6 WatchSource:0}: Error finding container a9cdc95bd12b4affa5977bc2df1cd089d7c82f225e74e0a87f1e3fce43707bb6: Status 404 returned error can't find the container with id a9cdc95bd12b4affa5977bc2df1cd089d7c82f225e74e0a87f1e3fce43707bb6 Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.470174 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dc4rc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-ddkhk_openstack-operators(4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.473793 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.474746 4979 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.474802 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert podName:3583aaeb-1aeb-4afc-91d2-a6837635a371 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:18.47478053 +0000 UTC m=+1102.749076334 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" (UID: "3583aaeb-1aeb-4afc-91d2-a6837635a371") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.479541 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dc4rc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-ddkhk_openstack-operators(4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.481512 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" podUID="4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04" Dec 04 12:01:16 crc kubenswrapper[4979]: W1204 12:01:16.485919 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa56245e_7ded_4797_9fd9_bc84274407e2.slice/crio-b465681c0463c0b185c75862db5b13bed311ebdc880e3a5cc2118a76ec1ab589 WatchSource:0}: Error finding container b465681c0463c0b185c75862db5b13bed311ebdc880e3a5cc2118a76ec1ab589: Status 404 returned error can't find the container with id b465681c0463c0b185c75862db5b13bed311ebdc880e3a5cc2118a76ec1ab589 Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.488325 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v"] Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.488799 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sd5sb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-c9c8v_openstack-operators(fa56245e-7ded-4797-9fd9-bc84274407e2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.491488 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sd5sb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-c9c8v_openstack-operators(fa56245e-7ded-4797-9fd9-bc84274407e2): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.494124 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" podUID="fa56245e-7ded-4797-9fd9-bc84274407e2" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.781746 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:16 crc kubenswrapper[4979]: I1204 12:01:16.782073 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.782214 4979 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.782257 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:18.782243039 +0000 UTC m=+1103.056538843 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "metrics-server-cert" not found Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.782639 4979 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 12:01:16 crc kubenswrapper[4979]: E1204 12:01:16.782666 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:18.78265926 +0000 UTC m=+1103.056955064 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "webhook-server-cert" not found Dec 04 12:01:17 crc kubenswrapper[4979]: I1204 12:01:17.324929 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" event={"ID":"4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04","Type":"ContainerStarted","Data":"a9cdc95bd12b4affa5977bc2df1cd089d7c82f225e74e0a87f1e3fce43707bb6"} Dec 04 12:01:17 crc kubenswrapper[4979]: E1204 12:01:17.336200 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" podUID="4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04" Dec 04 12:01:17 crc kubenswrapper[4979]: I1204 12:01:17.338339 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" event={"ID":"e5a502b0-c2bf-43fb-87eb-8def7494a0b9","Type":"ContainerStarted","Data":"efe12dc35f79bf9fcec65c1e88bee1530a0ef1111ca29d3808d99c42cb4a47b2"} Dec 04 12:01:17 crc kubenswrapper[4979]: I1204 12:01:17.340136 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" event={"ID":"d8e79a30-90c1-409c-a12b-ad2d0e0d36c8","Type":"ContainerStarted","Data":"eeb5c01030cd2f63ba0c9b8978012177b79b96dd362e2b18f636c7257f370fdb"} Dec 04 12:01:17 crc kubenswrapper[4979]: E1204 12:01:17.342024 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" podUID="d8e79a30-90c1-409c-a12b-ad2d0e0d36c8" Dec 04 12:01:17 crc kubenswrapper[4979]: I1204 12:01:17.354153 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" event={"ID":"91e3f7ac-b8d9-40d8-8a1c-94412c0a9488","Type":"ContainerStarted","Data":"26c963c7c38f0a854cdf3219f2d862ddf245f82365e4b5c81767b99d92d5deca"} Dec 04 12:01:17 crc kubenswrapper[4979]: E1204 12:01:17.356407 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" podUID="91e3f7ac-b8d9-40d8-8a1c-94412c0a9488" Dec 04 12:01:17 crc kubenswrapper[4979]: I1204 12:01:17.356396 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" event={"ID":"fa56245e-7ded-4797-9fd9-bc84274407e2","Type":"ContainerStarted","Data":"b465681c0463c0b185c75862db5b13bed311ebdc880e3a5cc2118a76ec1ab589"} Dec 04 12:01:17 crc kubenswrapper[4979]: E1204 12:01:17.359358 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" podUID="30066915-9744-4901-82c8-d19d06b1fb49" Dec 04 12:01:17 crc kubenswrapper[4979]: E1204 12:01:17.360563 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" podUID="fa56245e-7ded-4797-9fd9-bc84274407e2" Dec 04 12:01:18 crc kubenswrapper[4979]: I1204 12:01:18.008676 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.009114 4979 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.009353 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert podName:11b2439d-9796-4f42-94b4-f0903aa555d6 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:22.00933195 +0000 UTC m=+1106.283627754 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert") pod "infra-operator-controller-manager-57548d458d-2pnbm" (UID: "11b2439d-9796-4f42-94b4-f0903aa555d6") : secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.390281 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" podUID="4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04" Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.390503 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:7d66757c0af67104f0389e851a7cc0daa44443ad202d157417bd86bbb57cc385\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" podUID="91e3f7ac-b8d9-40d8-8a1c-94412c0a9488" Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.394508 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" podUID="d8e79a30-90c1-409c-a12b-ad2d0e0d36c8" Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.404907 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" podUID="fa56245e-7ded-4797-9fd9-bc84274407e2" Dec 04 12:01:18 crc kubenswrapper[4979]: I1204 12:01:18.516731 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.518906 4979 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.519012 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert podName:3583aaeb-1aeb-4afc-91d2-a6837635a371 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:22.51898952 +0000 UTC m=+1106.793285324 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" (UID: "3583aaeb-1aeb-4afc-91d2-a6837635a371") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: I1204 12:01:18.820457 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:18 crc kubenswrapper[4979]: I1204 12:01:18.820592 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.820761 4979 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.820821 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:22.820803754 +0000 UTC m=+1107.095099558 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "webhook-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.821192 4979 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 12:01:18 crc kubenswrapper[4979]: E1204 12:01:18.821230 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:22.821220855 +0000 UTC m=+1107.095516659 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "metrics-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: I1204 12:01:22.071770 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.071992 4979 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.072435 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert podName:11b2439d-9796-4f42-94b4-f0903aa555d6 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:30.072411667 +0000 UTC m=+1114.346707471 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert") pod "infra-operator-controller-manager-57548d458d-2pnbm" (UID: "11b2439d-9796-4f42-94b4-f0903aa555d6") : secret "infra-operator-webhook-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: I1204 12:01:22.580766 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.580951 4979 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.581029 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert podName:3583aaeb-1aeb-4afc-91d2-a6837635a371 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:30.581012799 +0000 UTC m=+1114.855308603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" (UID: "3583aaeb-1aeb-4afc-91d2-a6837635a371") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: I1204 12:01:22.885500 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:22 crc kubenswrapper[4979]: I1204 12:01:22.885640 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.885684 4979 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.885769 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:30.885749563 +0000 UTC m=+1115.160045377 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "metrics-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.885792 4979 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 12:01:22 crc kubenswrapper[4979]: E1204 12:01:22.885873 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:30.885855236 +0000 UTC m=+1115.160151120 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "webhook-server-cert" not found Dec 04 12:01:28 crc kubenswrapper[4979]: I1204 12:01:28.040887 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:01:28 crc kubenswrapper[4979]: I1204 12:01:28.041497 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:01:29 crc kubenswrapper[4979]: E1204 12:01:29.602289 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 04 12:01:29 crc kubenswrapper[4979]: E1204 12:01:29.602500 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6jwr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-kw2fk_openstack-operators(3369e6a8-2d6d-4c72-b96a-2cf650c4c257): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:01:30 crc kubenswrapper[4979]: E1204 12:01:30.076106 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 04 12:01:30 crc kubenswrapper[4979]: E1204 12:01:30.076515 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhwcg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-q2kv7_openstack-operators(0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.090372 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.102655 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/11b2439d-9796-4f42-94b4-f0903aa555d6-cert\") pod \"infra-operator-controller-manager-57548d458d-2pnbm\" (UID: \"11b2439d-9796-4f42-94b4-f0903aa555d6\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.144854 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.598355 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.601898 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/3583aaeb-1aeb-4afc-91d2-a6837635a371-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4xr262\" (UID: \"3583aaeb-1aeb-4afc-91d2-a6837635a371\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:30 crc kubenswrapper[4979]: E1204 12:01:30.628839 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 04 12:01:30 crc kubenswrapper[4979]: E1204 12:01:30.629068 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xt74s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-hgslt_openstack-operators(01d00bed-f29a-4ada-98d1-18ac5935eb06): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.744275 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.902060 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.902180 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:30 crc kubenswrapper[4979]: E1204 12:01:30.902253 4979 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 04 12:01:30 crc kubenswrapper[4979]: E1204 12:01:30.902347 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs podName:8ef350ec-4178-410b-9e8e-9981810dc148 nodeName:}" failed. No retries permitted until 2025-12-04 12:01:46.902327843 +0000 UTC m=+1131.176623647 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs") pod "openstack-operator-controller-manager-79cddf9bf6-xlr4l" (UID: "8ef350ec-4178-410b-9e8e-9981810dc148") : secret "webhook-server-cert" not found Dec 04 12:01:30 crc kubenswrapper[4979]: I1204 12:01:30.905475 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-metrics-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:39 crc kubenswrapper[4979]: I1204 12:01:39.711282 4979 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-fnchd container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": context deadline exceeded" start-of-body= Dec 04 12:01:39 crc kubenswrapper[4979]: I1204 12:01:39.711797 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" podUID="bb732274-6b17-47f6-ace6-485706ecd538" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": context deadline exceeded" Dec 04 12:01:39 crc kubenswrapper[4979]: I1204 12:01:39.712624 4979 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-fnchd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 12:01:39 crc kubenswrapper[4979]: I1204 12:01:39.712709 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" podUID="bb732274-6b17-47f6-ace6-485706ecd538" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 12:01:40 crc kubenswrapper[4979]: E1204 12:01:40.946544 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 04 12:01:40 crc kubenswrapper[4979]: E1204 12:01:40.946791 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-87wlp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-xwtfq_openstack-operators(4d6d4d26-72f0-4b63-979c-a533eb0dd53f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:01:41 crc kubenswrapper[4979]: E1204 12:01:41.454724 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 04 12:01:41 crc kubenswrapper[4979]: E1204 12:01:41.455775 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rf488,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-kk4p2_openstack-operators(e5a502b0-c2bf-43fb-87eb-8def7494a0b9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:01:41 crc kubenswrapper[4979]: E1204 12:01:41.457566 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" podUID="e5a502b0-c2bf-43fb-87eb-8def7494a0b9" Dec 04 12:01:41 crc kubenswrapper[4979]: I1204 12:01:41.681078 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm"] Dec 04 12:01:41 crc kubenswrapper[4979]: E1204 12:01:41.761018 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" podUID="e5a502b0-c2bf-43fb-87eb-8def7494a0b9" Dec 04 12:01:41 crc kubenswrapper[4979]: I1204 12:01:41.963806 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262"] Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.766483 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" event={"ID":"7c730311-2e6a-4815-87f4-f8099d5ced63","Type":"ContainerStarted","Data":"acf8c3066f7d174c6d21196e376ca1a2030129adfa5c67ff8cb00a52ae8b3eac"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.767512 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" event={"ID":"fe9dded8-cafd-469a-b65d-cfa21673e4b1","Type":"ContainerStarted","Data":"918509029aa6ceecd7aaf4ea393d9c2174d2e8b3700b930359ac7d7a814730da"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.768458 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" event={"ID":"5190c637-8327-4a8b-b351-4a6e86f6d59a","Type":"ContainerStarted","Data":"412074c0c334d5cf9130765a61a482cfb47268ff5ae06cf571e9ceaa6c212b02"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.771546 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" event={"ID":"d7efdf6b-7859-4750-b531-8bae5f2332fd","Type":"ContainerStarted","Data":"59c4d35cdd51d7abcf5165360c98fd01d0b09ad6fe9eaa01d203854496232d31"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.779324 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" event={"ID":"11b2439d-9796-4f42-94b4-f0903aa555d6","Type":"ContainerStarted","Data":"ad1ffc93479b29f8fa193e0696647068d51114d689cabc8aed9967bfe7193863"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.780773 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" event={"ID":"bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6","Type":"ContainerStarted","Data":"d6012b25aabd328546774685f22d053357d40e9052ebddb3d7d909a135073b09"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.783933 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" event={"ID":"e729014a-4926-4219-9159-785c4f602117","Type":"ContainerStarted","Data":"29ce72a94bc16b8172590b11d7fa18c801b66db8c066028d97cbdf25a58374af"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.795942 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" event={"ID":"465158ac-c7e8-4df8-97ee-231ba9ca3541","Type":"ContainerStarted","Data":"67f94d937dce234bb7f9549d9b59a501ac00c633b515278fcb6cdec0740a1339"} Dec 04 12:01:42 crc kubenswrapper[4979]: I1204 12:01:42.802247 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" event={"ID":"316f3ead-f9a1-44c5-84f9-5d8304a9f15b","Type":"ContainerStarted","Data":"30ed54f654ac117723502f2698fc7b2669d05229722bbb3d4d96b80aa0cf34ee"} Dec 04 12:01:43 crc kubenswrapper[4979]: I1204 12:01:43.812998 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" event={"ID":"3583aaeb-1aeb-4afc-91d2-a6837635a371","Type":"ContainerStarted","Data":"1ef95055ad896e2ef637eadd5f3d7214eaabf30047d716c43a6a141930b9b314"} Dec 04 12:01:43 crc kubenswrapper[4979]: I1204 12:01:43.814658 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" event={"ID":"cc4f8bbc-9399-4c17-a7a9-ec39a429813c","Type":"ContainerStarted","Data":"2a6112b6e89138bf488445cc95add2f9c072b64473d3917c0d464f4c500c588f"} Dec 04 12:01:43 crc kubenswrapper[4979]: I1204 12:01:43.816046 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" event={"ID":"aa22fecb-5237-48bd-9526-857132083b7f","Type":"ContainerStarted","Data":"6b6e96afda27b358bd1999a7701612ea5539f3fa1d06c82889fac22cfe8ccddc"} Dec 04 12:01:45 crc kubenswrapper[4979]: E1204 12:01:45.017490 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 04 12:01:45 crc kubenswrapper[4979]: E1204 12:01:45.017664 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6jwr2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-kw2fk_openstack-operators(3369e6a8-2d6d-4c72-b96a-2cf650c4c257): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 04 12:01:45 crc kubenswrapper[4979]: E1204 12:01:45.019399 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" podUID="3369e6a8-2d6d-4c72-b96a-2cf650c4c257" Dec 04 12:01:45 crc kubenswrapper[4979]: I1204 12:01:45.831808 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:01:46 crc kubenswrapper[4979]: I1204 12:01:46.928757 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:46 crc kubenswrapper[4979]: I1204 12:01:46.943226 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8ef350ec-4178-410b-9e8e-9981810dc148-webhook-certs\") pod \"openstack-operator-controller-manager-79cddf9bf6-xlr4l\" (UID: \"8ef350ec-4178-410b-9e8e-9981810dc148\") " pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:47 crc kubenswrapper[4979]: I1204 12:01:47.183057 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:48 crc kubenswrapper[4979]: I1204 12:01:48.535946 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l"] Dec 04 12:01:48 crc kubenswrapper[4979]: W1204 12:01:48.875602 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ef350ec_4178_410b_9e8e_9981810dc148.slice/crio-c8806053fef6ef4b5c24d937e4a2adebaea98574a96b1d7c8d11792ac2b6be92 WatchSource:0}: Error finding container c8806053fef6ef4b5c24d937e4a2adebaea98574a96b1d7c8d11792ac2b6be92: Status 404 returned error can't find the container with id c8806053fef6ef4b5c24d937e4a2adebaea98574a96b1d7c8d11792ac2b6be92 Dec 04 12:01:49 crc kubenswrapper[4979]: E1204 12:01:49.259707 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" podUID="01d00bed-f29a-4ada-98d1-18ac5935eb06" Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.869892 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" event={"ID":"3583aaeb-1aeb-4afc-91d2-a6837635a371","Type":"ContainerStarted","Data":"2a8e3959cf459d74ada443dba8ff2dd113e3f4837145c3c52dcbae8f933cc76f"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.904841 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" event={"ID":"11b2439d-9796-4f42-94b4-f0903aa555d6","Type":"ContainerStarted","Data":"0325afa7568ecfc6fb1754878721baad5ab1c8063277fe8d9ef2acc59b4048c3"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.904891 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" event={"ID":"11b2439d-9796-4f42-94b4-f0903aa555d6","Type":"ContainerStarted","Data":"fb55543bc3a5788caea1f6b939103c258ec1dd05734fa79f31fb2625e4fd9069"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.905058 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.926748 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" event={"ID":"30066915-9744-4901-82c8-d19d06b1fb49","Type":"ContainerStarted","Data":"5f1fd11ebaecec981aa28c1b0dde3e750cd00e7f2cafae2804a2a1bfc6e9f435"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.926802 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" event={"ID":"30066915-9744-4901-82c8-d19d06b1fb49","Type":"ContainerStarted","Data":"ed70db2a92109a20d36f0003db774c4398751cd314d0f6b975bf017896c92451"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.926986 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.930946 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" event={"ID":"01d00bed-f29a-4ada-98d1-18ac5935eb06","Type":"ContainerStarted","Data":"1a2b96025868ed52305f4311e3fa26585309fe9c3995bbf8761496943bef1af1"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.935871 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" event={"ID":"8ef350ec-4178-410b-9e8e-9981810dc148","Type":"ContainerStarted","Data":"8a099f3010d3602985c5563f3a0310a1b5ee032541d1741db0356fb17784d039"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.935926 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.935942 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" event={"ID":"8ef350ec-4178-410b-9e8e-9981810dc148","Type":"ContainerStarted","Data":"c8806053fef6ef4b5c24d937e4a2adebaea98574a96b1d7c8d11792ac2b6be92"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.963372 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" event={"ID":"4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04","Type":"ContainerStarted","Data":"f5481096a2eee65bcb0e66970cd9b3cb86a5b745a843098f75c94abba3971e9b"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.965802 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" event={"ID":"3369e6a8-2d6d-4c72-b96a-2cf650c4c257","Type":"ContainerStarted","Data":"eab91e236d5fde0b998ab775708304776099e6b9383d55369bb9a21819b262c3"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.968379 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" event={"ID":"d8e79a30-90c1-409c-a12b-ad2d0e0d36c8","Type":"ContainerStarted","Data":"04ced0899215999eb50dcb1171ee124a4846ad0ca931fb4f35b360931f323b65"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.968421 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" event={"ID":"d8e79a30-90c1-409c-a12b-ad2d0e0d36c8","Type":"ContainerStarted","Data":"430b7dc31e87522af687a73d67f06d5295faf37b8f2485631dcf77346a467d5e"} Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.969197 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.985249 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" podStartSLOduration=29.992809046 podStartE2EDuration="35.985228413s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:41.761117967 +0000 UTC m=+1126.035413771" lastFinishedPulling="2025-12-04 12:01:47.753537334 +0000 UTC m=+1132.027833138" observedRunningTime="2025-12-04 12:01:49.935505989 +0000 UTC m=+1134.209801913" watchObservedRunningTime="2025-12-04 12:01:49.985228413 +0000 UTC m=+1134.259524217" Dec 04 12:01:49 crc kubenswrapper[4979]: I1204 12:01:49.991548 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" podStartSLOduration=35.991525725 podStartE2EDuration="35.991525725s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:01:49.986080637 +0000 UTC m=+1134.260376441" watchObservedRunningTime="2025-12-04 12:01:49.991525725 +0000 UTC m=+1134.265821529" Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.012901 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" event={"ID":"cc4f8bbc-9399-4c17-a7a9-ec39a429813c","Type":"ContainerStarted","Data":"e87bfdc617cc0c55debea5af154af0b017358201ec7bdc653a4e34e8aa3c0984"} Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.015065 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.027612 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.040715 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" event={"ID":"91e3f7ac-b8d9-40d8-8a1c-94412c0a9488","Type":"ContainerStarted","Data":"94b47578b2e7a1ca0201b36becff641629490bf5387d3237ece3ceac042ac148"} Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.048105 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" event={"ID":"fa56245e-7ded-4797-9fd9-bc84274407e2","Type":"ContainerStarted","Data":"bd1481acec65585afa53b3ed4e210562ea374446eeb904a57ca3c1ebb1201bd9"} Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.073483 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" podStartSLOduration=4.95639119 podStartE2EDuration="36.073464878s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.106695859 +0000 UTC m=+1100.380991663" lastFinishedPulling="2025-12-04 12:01:47.223769547 +0000 UTC m=+1131.498065351" observedRunningTime="2025-12-04 12:01:50.040624663 +0000 UTC m=+1134.314920467" watchObservedRunningTime="2025-12-04 12:01:50.073464878 +0000 UTC m=+1134.347760672" Dec 04 12:01:50 crc kubenswrapper[4979]: I1204 12:01:50.098493 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" podStartSLOduration=5.289925608 podStartE2EDuration="36.098476849s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.388718874 +0000 UTC m=+1100.663014678" lastFinishedPulling="2025-12-04 12:01:47.197270115 +0000 UTC m=+1131.471565919" observedRunningTime="2025-12-04 12:01:50.064536845 +0000 UTC m=+1134.338832659" watchObservedRunningTime="2025-12-04 12:01:50.098476849 +0000 UTC m=+1134.372772643" Dec 04 12:01:50 crc kubenswrapper[4979]: E1204 12:01:50.414978 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" podUID="4d6d4d26-72f0-4b63-979c-a533eb0dd53f" Dec 04 12:01:50 crc kubenswrapper[4979]: E1204 12:01:50.495205 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" podUID="0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.057322 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" event={"ID":"4d6d4d26-72f0-4b63-979c-a533eb0dd53f","Type":"ContainerStarted","Data":"46a62102ad887f0e0581b6140a1f902d811a2b2fb9c28dec67074538eedc395b"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.059959 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" event={"ID":"4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04","Type":"ContainerStarted","Data":"995a6b4f9f0971cba20de895462416140c05ba338af9cc862e06f21f98003a35"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.060324 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.062005 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" event={"ID":"5190c637-8327-4a8b-b351-4a6e86f6d59a","Type":"ContainerStarted","Data":"bc6738797b6031f5d801ad5c6e98d6d1c73859bf1ac73fa82284980000f89ef8"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.062283 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.063949 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.064585 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" event={"ID":"3369e6a8-2d6d-4c72-b96a-2cf650c4c257","Type":"ContainerStarted","Data":"c7ec864e0c78be5f1324e000900638449cc2ecbd978bdc2c3e54f2303d6e59c7"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.064774 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.066060 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" event={"ID":"0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b","Type":"ContainerStarted","Data":"7b4e74345776eb80b53085e86107b73fc66c178eb227c6d837a5f591d8cf093e"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.067829 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" event={"ID":"bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6","Type":"ContainerStarted","Data":"32281b8b26befadabb5695fd6b74d183ce22103313080333e389c1c334e7879d"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.068087 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.069378 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" event={"ID":"7c730311-2e6a-4815-87f4-f8099d5ced63","Type":"ContainerStarted","Data":"fd2743e370d4268babf32f4be46fb3275cd755eb3cbe325e576353283a47ce13"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.069868 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.074944 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" event={"ID":"fa56245e-7ded-4797-9fd9-bc84274407e2","Type":"ContainerStarted","Data":"d03f7d5f88205e2a57039d4965d2b363634f30fea5055b53932b797f681e2078"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.075080 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.075132 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.076533 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.076976 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" event={"ID":"316f3ead-f9a1-44c5-84f9-5d8304a9f15b","Type":"ContainerStarted","Data":"24f650a33be87ed2a6a9711f379203faeffbbb9b7c6942e87a60f62b0a3c1558"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.078966 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.080704 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.080726 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" event={"ID":"aa22fecb-5237-48bd-9526-857132083b7f","Type":"ContainerStarted","Data":"dbdd0497d26c3a1dfe4973839f5bac1c21929c58f7f798c4da6475c52dc2698c"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.080895 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.084060 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" event={"ID":"e729014a-4926-4219-9159-785c4f602117","Type":"ContainerStarted","Data":"dff93f4eabedeb289154e2224a18d0eb9a3fe5d67cd8c457241250c14bd6f013"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.084240 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.085960 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" event={"ID":"465158ac-c7e8-4df8-97ee-231ba9ca3541","Type":"ContainerStarted","Data":"1723d1d602120c06345f2b45d59a7c1b98193f1c7cb736a1a376ed3a3cf3b928"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.086165 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.087989 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.089073 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" event={"ID":"01d00bed-f29a-4ada-98d1-18ac5935eb06","Type":"ContainerStarted","Data":"43f45c6f40096a06e3d4ee42bfbd7608575ece458563e0a22df7a29d876b1889"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.089393 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.090548 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.091072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" event={"ID":"d7efdf6b-7859-4750-b531-8bae5f2332fd","Type":"ContainerStarted","Data":"d59fabf9729b10740a7ba0ff1f35070c81a8b3b6e6c807120e46a8bd2c680d45"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.091719 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.093315 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" event={"ID":"3583aaeb-1aeb-4afc-91d2-a6837635a371","Type":"ContainerStarted","Data":"fc9c57c3990a0ae8074cb391b26bda0810cb9fdabdfe10be3abbc1d0b2d458d7"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.093500 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.094163 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-5vbf8" podStartSLOduration=3.108323076 podStartE2EDuration="37.094152214s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.115536888 +0000 UTC m=+1099.389832692" lastFinishedPulling="2025-12-04 12:01:49.101366026 +0000 UTC m=+1133.375661830" observedRunningTime="2025-12-04 12:01:50.102429577 +0000 UTC m=+1134.376725381" watchObservedRunningTime="2025-12-04 12:01:51.094152214 +0000 UTC m=+1135.368448018" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.095805 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" event={"ID":"91e3f7ac-b8d9-40d8-8a1c-94412c0a9488","Type":"ContainerStarted","Data":"9e5381146867018a08e7053d7cb43cb523e291c58e277827379c67918ce34e1a"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.096206 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.098525 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" event={"ID":"fe9dded8-cafd-469a-b65d-cfa21673e4b1","Type":"ContainerStarted","Data":"fd7415367b9bb0e256d6df7d95fd412d056467b294d8c350cea13b2863b30475"} Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.098551 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.107728 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.110438 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.130842 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.184804 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7gb6r" podStartSLOduration=2.972277368 podStartE2EDuration="37.184787954s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.974948199 +0000 UTC m=+1100.249244003" lastFinishedPulling="2025-12-04 12:01:50.187458785 +0000 UTC m=+1134.461754589" observedRunningTime="2025-12-04 12:01:51.141116794 +0000 UTC m=+1135.415412598" watchObservedRunningTime="2025-12-04 12:01:51.184787954 +0000 UTC m=+1135.459083758" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.219907 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" podStartSLOduration=6.466205586 podStartE2EDuration="37.219884801s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.469979799 +0000 UTC m=+1100.744275603" lastFinishedPulling="2025-12-04 12:01:47.223659014 +0000 UTC m=+1131.497954818" observedRunningTime="2025-12-04 12:01:51.212965772 +0000 UTC m=+1135.487261596" watchObservedRunningTime="2025-12-04 12:01:51.219884801 +0000 UTC m=+1135.494180605" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.262215 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" podStartSLOduration=4.877855979 podStartE2EDuration="37.262179333s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.520190505 +0000 UTC m=+1099.794486309" lastFinishedPulling="2025-12-04 12:01:47.904513859 +0000 UTC m=+1132.178809663" observedRunningTime="2025-12-04 12:01:51.261634068 +0000 UTC m=+1135.535929872" watchObservedRunningTime="2025-12-04 12:01:51.262179333 +0000 UTC m=+1135.536475137" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.262542 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-lw7hk" podStartSLOduration=2.586437632 podStartE2EDuration="37.262538402s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.616263193 +0000 UTC m=+1099.890558997" lastFinishedPulling="2025-12-04 12:01:50.292363963 +0000 UTC m=+1134.566659767" observedRunningTime="2025-12-04 12:01:51.243698379 +0000 UTC m=+1135.517994183" watchObservedRunningTime="2025-12-04 12:01:51.262538402 +0000 UTC m=+1135.536834206" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.294783 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xgjxw" podStartSLOduration=3.818958642 podStartE2EDuration="37.294765541s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.205752448 +0000 UTC m=+1100.480048252" lastFinishedPulling="2025-12-04 12:01:49.681559347 +0000 UTC m=+1133.955855151" observedRunningTime="2025-12-04 12:01:51.291629015 +0000 UTC m=+1135.565924819" watchObservedRunningTime="2025-12-04 12:01:51.294765541 +0000 UTC m=+1135.569061345" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.321332 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" podStartSLOduration=6.056377236 podStartE2EDuration="37.321312774s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.488653678 +0000 UTC m=+1100.762949482" lastFinishedPulling="2025-12-04 12:01:47.753589216 +0000 UTC m=+1132.027885020" observedRunningTime="2025-12-04 12:01:51.313933543 +0000 UTC m=+1135.588229347" watchObservedRunningTime="2025-12-04 12:01:51.321312774 +0000 UTC m=+1135.595608578" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.426996 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-mr2rj" podStartSLOduration=3.636740286 podStartE2EDuration="37.426965024s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.61615064 +0000 UTC m=+1099.890446444" lastFinishedPulling="2025-12-04 12:01:49.406375378 +0000 UTC m=+1133.680671182" observedRunningTime="2025-12-04 12:01:51.356592236 +0000 UTC m=+1135.630888030" watchObservedRunningTime="2025-12-04 12:01:51.426965024 +0000 UTC m=+1135.701260828" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.472026 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5rcqm" podStartSLOduration=2.81313677 podStartE2EDuration="37.465289458s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.782978536 +0000 UTC m=+1100.057274340" lastFinishedPulling="2025-12-04 12:01:50.435131234 +0000 UTC m=+1134.709427028" observedRunningTime="2025-12-04 12:01:51.403284879 +0000 UTC m=+1135.677580693" watchObservedRunningTime="2025-12-04 12:01:51.465289458 +0000 UTC m=+1135.739585252" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.472378 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" podStartSLOduration=3.400443547 podStartE2EDuration="37.472372541s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.102547156 +0000 UTC m=+1100.376842960" lastFinishedPulling="2025-12-04 12:01:50.17447615 +0000 UTC m=+1134.448771954" observedRunningTime="2025-12-04 12:01:51.447837943 +0000 UTC m=+1135.722133747" watchObservedRunningTime="2025-12-04 12:01:51.472372541 +0000 UTC m=+1135.746668345" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.486541 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-xzkn7" podStartSLOduration=4.283903592 podStartE2EDuration="37.486518646s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.061580189 +0000 UTC m=+1100.335875993" lastFinishedPulling="2025-12-04 12:01:49.264195243 +0000 UTC m=+1133.538491047" observedRunningTime="2025-12-04 12:01:51.472163345 +0000 UTC m=+1135.746459149" watchObservedRunningTime="2025-12-04 12:01:51.486518646 +0000 UTC m=+1135.760814450" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.548645 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" podStartSLOduration=6.482289825 podStartE2EDuration="37.54862623s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.214452796 +0000 UTC m=+1100.488748600" lastFinishedPulling="2025-12-04 12:01:47.280789201 +0000 UTC m=+1131.555085005" observedRunningTime="2025-12-04 12:01:51.541865775 +0000 UTC m=+1135.816161579" watchObservedRunningTime="2025-12-04 12:01:51.54862623 +0000 UTC m=+1135.822922024" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.552878 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-hwjs2" podStartSLOduration=3.80166163 podStartE2EDuration="37.552860475s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.969073268 +0000 UTC m=+1100.243369072" lastFinishedPulling="2025-12-04 12:01:49.720272113 +0000 UTC m=+1133.994567917" observedRunningTime="2025-12-04 12:01:51.504592109 +0000 UTC m=+1135.778887913" watchObservedRunningTime="2025-12-04 12:01:51.552860475 +0000 UTC m=+1135.827156279" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.585655 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-tv9lc" podStartSLOduration=2.815790313 podStartE2EDuration="37.585635948s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.698268968 +0000 UTC m=+1099.972564772" lastFinishedPulling="2025-12-04 12:01:50.468114603 +0000 UTC m=+1134.742410407" observedRunningTime="2025-12-04 12:01:51.580476297 +0000 UTC m=+1135.854772101" watchObservedRunningTime="2025-12-04 12:01:51.585635948 +0000 UTC m=+1135.859931742" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.610885 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" podStartSLOduration=2.5183329370000003 podStartE2EDuration="37.610868996s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.519607499 +0000 UTC m=+1099.793903313" lastFinishedPulling="2025-12-04 12:01:50.612143568 +0000 UTC m=+1134.886439372" observedRunningTime="2025-12-04 12:01:51.604552544 +0000 UTC m=+1135.878848348" watchObservedRunningTime="2025-12-04 12:01:51.610868996 +0000 UTC m=+1135.885164800" Dec 04 12:01:51 crc kubenswrapper[4979]: I1204 12:01:51.644776 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" podStartSLOduration=33.171247996 podStartE2EDuration="37.644758009s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:43.428449076 +0000 UTC m=+1127.702744880" lastFinishedPulling="2025-12-04 12:01:47.901959089 +0000 UTC m=+1132.176254893" observedRunningTime="2025-12-04 12:01:51.627616642 +0000 UTC m=+1135.901912446" watchObservedRunningTime="2025-12-04 12:01:51.644758009 +0000 UTC m=+1135.919053823" Dec 04 12:01:52 crc kubenswrapper[4979]: I1204 12:01:52.105787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" event={"ID":"0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b","Type":"ContainerStarted","Data":"24698783dadb827e89f568e719d3949464180c607fd0c55f9dbe16ed5f7ac87b"} Dec 04 12:01:52 crc kubenswrapper[4979]: I1204 12:01:52.106143 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:01:52 crc kubenswrapper[4979]: I1204 12:01:52.108260 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" event={"ID":"4d6d4d26-72f0-4b63-979c-a533eb0dd53f","Type":"ContainerStarted","Data":"d5c22935a01bffd32a97cc104b91a8d179edb29cb1ce0db4b4203bd688c9aa6a"} Dec 04 12:01:52 crc kubenswrapper[4979]: I1204 12:01:52.127853 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" podStartSLOduration=2.416021429 podStartE2EDuration="38.127834924s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.910703268 +0000 UTC m=+1100.184999082" lastFinishedPulling="2025-12-04 12:01:51.622516773 +0000 UTC m=+1135.896812577" observedRunningTime="2025-12-04 12:01:52.12290925 +0000 UTC m=+1136.397205054" watchObservedRunningTime="2025-12-04 12:01:52.127834924 +0000 UTC m=+1136.402130728" Dec 04 12:01:52 crc kubenswrapper[4979]: I1204 12:01:52.148666 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" podStartSLOduration=2.515707664 podStartE2EDuration="38.148649321s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:15.983224983 +0000 UTC m=+1100.257520797" lastFinishedPulling="2025-12-04 12:01:51.61616665 +0000 UTC m=+1135.890462454" observedRunningTime="2025-12-04 12:01:52.141450275 +0000 UTC m=+1136.415746089" watchObservedRunningTime="2025-12-04 12:01:52.148649321 +0000 UTC m=+1136.422945125" Dec 04 12:01:53 crc kubenswrapper[4979]: I1204 12:01:53.117071 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:01:54 crc kubenswrapper[4979]: I1204 12:01:54.448896 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-kw2fk" Dec 04 12:01:55 crc kubenswrapper[4979]: I1204 12:01:55.169976 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-7jr2r" Dec 04 12:01:55 crc kubenswrapper[4979]: I1204 12:01:55.232395 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-dhmqs" Dec 04 12:01:55 crc kubenswrapper[4979]: I1204 12:01:55.272869 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-55g6q" Dec 04 12:01:55 crc kubenswrapper[4979]: I1204 12:01:55.298416 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-c9c8v" Dec 04 12:01:55 crc kubenswrapper[4979]: I1204 12:01:55.326526 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-ddkhk" Dec 04 12:01:57 crc kubenswrapper[4979]: I1204 12:01:57.145869 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" event={"ID":"e5a502b0-c2bf-43fb-87eb-8def7494a0b9","Type":"ContainerStarted","Data":"5ce00222112372465e502b4be83b43c4a58bd5af7e81bf34d78f19b164bee2e2"} Dec 04 12:01:57 crc kubenswrapper[4979]: I1204 12:01:57.166654 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-kk4p2" podStartSLOduration=2.781846548 podStartE2EDuration="43.166627003s" podCreationTimestamp="2025-12-04 12:01:14 +0000 UTC" firstStartedPulling="2025-12-04 12:01:16.344478949 +0000 UTC m=+1100.618774753" lastFinishedPulling="2025-12-04 12:01:56.729259404 +0000 UTC m=+1141.003555208" observedRunningTime="2025-12-04 12:01:57.163269501 +0000 UTC m=+1141.437565315" watchObservedRunningTime="2025-12-04 12:01:57.166627003 +0000 UTC m=+1141.440922837" Dec 04 12:01:57 crc kubenswrapper[4979]: I1204 12:01:57.191099 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-79cddf9bf6-xlr4l" Dec 04 12:01:58 crc kubenswrapper[4979]: I1204 12:01:58.041283 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:01:58 crc kubenswrapper[4979]: I1204 12:01:58.041649 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:01:58 crc kubenswrapper[4979]: I1204 12:01:58.041696 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:01:58 crc kubenswrapper[4979]: I1204 12:01:58.042363 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb5b5fb8f2704900b621034dde12cd64efd7f6ba3338a4dd8572ae9e51dbc1d8"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:01:58 crc kubenswrapper[4979]: I1204 12:01:58.042421 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://bb5b5fb8f2704900b621034dde12cd64efd7f6ba3338a4dd8572ae9e51dbc1d8" gracePeriod=600 Dec 04 12:01:59 crc kubenswrapper[4979]: I1204 12:01:59.162672 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="bb5b5fb8f2704900b621034dde12cd64efd7f6ba3338a4dd8572ae9e51dbc1d8" exitCode=0 Dec 04 12:01:59 crc kubenswrapper[4979]: I1204 12:01:59.162757 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"bb5b5fb8f2704900b621034dde12cd64efd7f6ba3338a4dd8572ae9e51dbc1d8"} Dec 04 12:01:59 crc kubenswrapper[4979]: I1204 12:01:59.163063 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"fb52f2da72fff7d717ea5b0066df40b5dc5b046752472d6fe21e232348fdf595"} Dec 04 12:01:59 crc kubenswrapper[4979]: I1204 12:01:59.163110 4979 scope.go:117] "RemoveContainer" containerID="8eaf46a3c990101fdd147fc7b0c6c360a6dd66b15c629fa454f266dceff5343d" Dec 04 12:02:00 crc kubenswrapper[4979]: I1204 12:02:00.154900 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-2pnbm" Dec 04 12:02:00 crc kubenswrapper[4979]: I1204 12:02:00.751208 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4xr262" Dec 04 12:02:04 crc kubenswrapper[4979]: I1204 12:02:04.495467 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hgslt" Dec 04 12:02:04 crc kubenswrapper[4979]: I1204 12:02:04.718430 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-q2kv7" Dec 04 12:02:04 crc kubenswrapper[4979]: I1204 12:02:04.885871 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-xwtfq" Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.957386 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jj8xn"] Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.959217 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.962761 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.964450 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.964852 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-rcnxw" Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.975396 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 12:02:19 crc kubenswrapper[4979]: I1204 12:02:19.991935 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jj8xn"] Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.033412 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-config\") pod \"dnsmasq-dns-675f4bcbfc-jj8xn\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.033475 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dv8f\" (UniqueName: \"kubernetes.io/projected/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-kube-api-access-6dv8f\") pod \"dnsmasq-dns-675f4bcbfc-jj8xn\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.047442 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rm27j"] Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.049063 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.055378 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.055822 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rm27j"] Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.135143 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-config\") pod \"dnsmasq-dns-675f4bcbfc-jj8xn\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.135193 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dv8f\" (UniqueName: \"kubernetes.io/projected/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-kube-api-access-6dv8f\") pod \"dnsmasq-dns-675f4bcbfc-jj8xn\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.135268 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx6s2\" (UniqueName: \"kubernetes.io/projected/52fdc353-c66e-46d4-9021-869535523f3d-kube-api-access-vx6s2\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.135291 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.135419 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-config\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.136155 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-config\") pod \"dnsmasq-dns-675f4bcbfc-jj8xn\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.154115 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dv8f\" (UniqueName: \"kubernetes.io/projected/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-kube-api-access-6dv8f\") pod \"dnsmasq-dns-675f4bcbfc-jj8xn\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.236869 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx6s2\" (UniqueName: \"kubernetes.io/projected/52fdc353-c66e-46d4-9021-869535523f3d-kube-api-access-vx6s2\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.236921 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.236982 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-config\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.237925 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.237990 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-config\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.253378 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx6s2\" (UniqueName: \"kubernetes.io/projected/52fdc353-c66e-46d4-9021-869535523f3d-kube-api-access-vx6s2\") pod \"dnsmasq-dns-78dd6ddcc-rm27j\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.286172 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.368402 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.716315 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jj8xn"] Dec 04 12:02:20 crc kubenswrapper[4979]: I1204 12:02:20.806610 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rm27j"] Dec 04 12:02:20 crc kubenswrapper[4979]: W1204 12:02:20.807133 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod52fdc353_c66e_46d4_9021_869535523f3d.slice/crio-22a815c1dec64065e00120719ea5be2eabba2e7e5b64735c1259efdf42df8369 WatchSource:0}: Error finding container 22a815c1dec64065e00120719ea5be2eabba2e7e5b64735c1259efdf42df8369: Status 404 returned error can't find the container with id 22a815c1dec64065e00120719ea5be2eabba2e7e5b64735c1259efdf42df8369 Dec 04 12:02:21 crc kubenswrapper[4979]: I1204 12:02:21.334382 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" event={"ID":"52fdc353-c66e-46d4-9021-869535523f3d","Type":"ContainerStarted","Data":"22a815c1dec64065e00120719ea5be2eabba2e7e5b64735c1259efdf42df8369"} Dec 04 12:02:21 crc kubenswrapper[4979]: I1204 12:02:21.335913 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" event={"ID":"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc","Type":"ContainerStarted","Data":"dfe83af52541f0bfaae7fb5b9fe900bc443f1041742fa6c5616250c9e71b9ac1"} Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.787439 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jj8xn"] Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.832573 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kmnvq"] Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.838279 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.841189 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kmnvq"] Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.989156 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-config\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.989570 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:22 crc kubenswrapper[4979]: I1204 12:02:22.989769 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zn6b\" (UniqueName: \"kubernetes.io/projected/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-kube-api-access-5zn6b\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.087597 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rm27j"] Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.090904 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.091195 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zn6b\" (UniqueName: \"kubernetes.io/projected/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-kube-api-access-5zn6b\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.091310 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-config\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.091854 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.092070 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-config\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.121762 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zn6b\" (UniqueName: \"kubernetes.io/projected/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-kube-api-access-5zn6b\") pod \"dnsmasq-dns-666b6646f7-kmnvq\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.128970 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p4zbn"] Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.130176 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.135330 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p4zbn"] Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.171529 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.193555 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-config\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.193591 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.193628 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqqh8\" (UniqueName: \"kubernetes.io/projected/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-kube-api-access-nqqh8\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.294995 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-config\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.295038 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.295114 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqqh8\" (UniqueName: \"kubernetes.io/projected/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-kube-api-access-nqqh8\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.296731 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.297234 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-config\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.317952 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqqh8\" (UniqueName: \"kubernetes.io/projected/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-kube-api-access-nqqh8\") pod \"dnsmasq-dns-57d769cc4f-p4zbn\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.471773 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.647611 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kmnvq"] Dec 04 12:02:23 crc kubenswrapper[4979]: W1204 12:02:23.660620 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9a2aa4b_cea0_40e4_94b8_d19edadde1e7.slice/crio-b183cdfe9e048ae2c7fc460930178e0df568919de7c2a9406d0ea18085f61373 WatchSource:0}: Error finding container b183cdfe9e048ae2c7fc460930178e0df568919de7c2a9406d0ea18085f61373: Status 404 returned error can't find the container with id b183cdfe9e048ae2c7fc460930178e0df568919de7c2a9406d0ea18085f61373 Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.951786 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.953833 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.958156 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.960097 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-8tpw8" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.960100 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.960148 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.960205 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.960386 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.960414 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.977942 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p4zbn"] Dec 04 12:02:23 crc kubenswrapper[4979]: W1204 12:02:23.986884 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c1e3480_dc4a_438c_a6d0_ede582da1a7e.slice/crio-e7c59f7abed59d9f5ff4c6889e2d2cc265d46fbdb0a7be490c19f85650c8ee01 WatchSource:0}: Error finding container e7c59f7abed59d9f5ff4c6889e2d2cc265d46fbdb0a7be490c19f85650c8ee01: Status 404 returned error can't find the container with id e7c59f7abed59d9f5ff4c6889e2d2cc265d46fbdb0a7be490c19f85650c8ee01 Dec 04 12:02:23 crc kubenswrapper[4979]: I1204 12:02:23.987358 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.112839 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.112903 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzm4n\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-kube-api-access-jzm4n\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113099 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113157 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113186 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113212 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113373 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113430 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113533 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113781 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.113832 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.214856 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.214921 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.214969 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.214997 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzm4n\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-kube-api-access-jzm4n\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215034 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215064 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215084 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215113 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215146 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215179 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215211 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215524 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.215562 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.216037 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.216748 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.216754 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.217232 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-server-conf\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.224554 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.224864 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-pod-info\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.229987 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.237474 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.248265 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.252334 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.262700 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzm4n\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-kube-api-access-jzm4n\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.263357 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-gpcrd" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.263800 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.263910 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.263961 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.263978 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.263913 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.264098 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.272123 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.292837 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-server-0\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.316761 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.316891 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.316927 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.316957 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.316981 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.317038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.317186 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.317310 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0fad7195-d19a-48ce-ad3d-4c67b53d2974-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.317342 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0fad7195-d19a-48ce-ad3d-4c67b53d2974-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.317372 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.317393 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkr8v\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-kube-api-access-jkr8v\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.388914 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" event={"ID":"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7","Type":"ContainerStarted","Data":"b183cdfe9e048ae2c7fc460930178e0df568919de7c2a9406d0ea18085f61373"} Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.403117 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" event={"ID":"3c1e3480-dc4a-438c-a6d0-ede582da1a7e","Type":"ContainerStarted","Data":"e7c59f7abed59d9f5ff4c6889e2d2cc265d46fbdb0a7be490c19f85650c8ee01"} Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421384 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421448 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421468 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421488 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421504 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421548 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421571 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0fad7195-d19a-48ce-ad3d-4c67b53d2974-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421585 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0fad7195-d19a-48ce-ad3d-4c67b53d2974-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421602 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421618 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkr8v\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-kube-api-access-jkr8v\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.421652 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.422871 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.424144 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.424281 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.424525 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.424745 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.425429 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.431415 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.439988 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0fad7195-d19a-48ce-ad3d-4c67b53d2974-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.441099 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0fad7195-d19a-48ce-ad3d-4c67b53d2974-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.450949 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkr8v\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-kube-api-access-jkr8v\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.465645 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.470815 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.586545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 12:02:24 crc kubenswrapper[4979]: I1204 12:02:24.629274 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.125328 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.233813 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:02:25 crc kubenswrapper[4979]: W1204 12:02:25.248026 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fad7195_d19a_48ce_ad3d_4c67b53d2974.slice/crio-35f166aeb381070445583e255246ccc0b4489934f3c5c7566f44d3721006b0bb WatchSource:0}: Error finding container 35f166aeb381070445583e255246ccc0b4489934f3c5c7566f44d3721006b0bb: Status 404 returned error can't find the container with id 35f166aeb381070445583e255246ccc0b4489934f3c5c7566f44d3721006b0bb Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.414816 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0fad7195-d19a-48ce-ad3d-4c67b53d2974","Type":"ContainerStarted","Data":"35f166aeb381070445583e255246ccc0b4489934f3c5c7566f44d3721006b0bb"} Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.416501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48","Type":"ContainerStarted","Data":"8325f2ca802d79b896407aaa3d575bc1bb498938c81b1d45cb08ee25449c8f74"} Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.482339 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.484133 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.490077 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-7tqtd" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.490717 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.490832 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.490943 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.495033 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.502944 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.546841 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-kolla-config\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.546886 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.546937 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-config-data-default\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.546958 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.546978 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsc2q\" (UniqueName: \"kubernetes.io/projected/7130263f-15db-4a87-a078-9c346739eced-kube-api-access-jsc2q\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.547003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.547031 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7130263f-15db-4a87-a078-9c346739eced-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.547060 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.648991 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7130263f-15db-4a87-a078-9c346739eced-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649058 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649112 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-kolla-config\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649137 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649199 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-config-data-default\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649257 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsc2q\" (UniqueName: \"kubernetes.io/projected/7130263f-15db-4a87-a078-9c346739eced-kube-api-access-jsc2q\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649313 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.649629 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.653693 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-operator-scripts\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.653763 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7130263f-15db-4a87-a078-9c346739eced-config-data-generated\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.654371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-kolla-config\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.654757 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-config-data-default\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.657778 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.673390 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.684513 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.703934 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsc2q\" (UniqueName: \"kubernetes.io/projected/7130263f-15db-4a87-a078-9c346739eced-kube-api-access-jsc2q\") pod \"openstack-galera-0\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " pod="openstack/openstack-galera-0" Dec 04 12:02:25 crc kubenswrapper[4979]: I1204 12:02:25.809176 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.474568 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 12:02:26 crc kubenswrapper[4979]: W1204 12:02:26.494779 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7130263f_15db_4a87_a078_9c346739eced.slice/crio-ac4aaca30cc1606e8c5bf93c112be1c940e01ff1a56fdd42fc23c84df674475b WatchSource:0}: Error finding container ac4aaca30cc1606e8c5bf93c112be1c940e01ff1a56fdd42fc23c84df674475b: Status 404 returned error can't find the container with id ac4aaca30cc1606e8c5bf93c112be1c940e01ff1a56fdd42fc23c84df674475b Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.916502 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.917879 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.920465 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.920958 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.921394 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.922573 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-jw8kh" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.926572 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970453 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970507 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970543 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970570 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwx8t\" (UniqueName: \"kubernetes.io/projected/d3791d23-03e1-4c8d-b44e-738eff9222d0-kube-api-access-kwx8t\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970694 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970759 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.970915 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:26 crc kubenswrapper[4979]: I1204 12:02:26.971059 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072633 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072698 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072722 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwx8t\" (UniqueName: \"kubernetes.io/projected/d3791d23-03e1-4c8d-b44e-738eff9222d0-kube-api-access-kwx8t\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072737 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072756 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072792 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072830 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.072870 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.073647 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.074290 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.076437 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.076497 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.077173 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.092410 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.104954 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwx8t\" (UniqueName: \"kubernetes.io/projected/d3791d23-03e1-4c8d-b44e-738eff9222d0-kube-api-access-kwx8t\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.120111 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.120984 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.265568 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.311888 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.313593 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.320183 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.320392 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.320461 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-qpcxr" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.326764 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.382125 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.382182 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-kolla-config\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.382205 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.382470 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-config-data\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.384161 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wszst\" (UniqueName: \"kubernetes.io/projected/f763e763-10df-408d-9452-84634d47b455-kube-api-access-wszst\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.461328 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7130263f-15db-4a87-a078-9c346739eced","Type":"ContainerStarted","Data":"ac4aaca30cc1606e8c5bf93c112be1c940e01ff1a56fdd42fc23c84df674475b"} Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.486611 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.486668 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-kolla-config\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.487168 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.487221 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-config-data\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.488316 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wszst\" (UniqueName: \"kubernetes.io/projected/f763e763-10df-408d-9452-84634d47b455-kube-api-access-wszst\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.489803 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-kolla-config\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.492066 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-config-data\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.494966 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-memcached-tls-certs\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.507199 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wszst\" (UniqueName: \"kubernetes.io/projected/f763e763-10df-408d-9452-84634d47b455-kube-api-access-wszst\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.507798 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-combined-ca-bundle\") pod \"memcached-0\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.641384 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 12:02:27 crc kubenswrapper[4979]: I1204 12:02:27.934362 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 12:02:28 crc kubenswrapper[4979]: I1204 12:02:28.193387 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 12:02:28 crc kubenswrapper[4979]: I1204 12:02:28.881222 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:02:28 crc kubenswrapper[4979]: I1204 12:02:28.882134 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:02:28 crc kubenswrapper[4979]: I1204 12:02:28.888403 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-wsqqr" Dec 04 12:02:28 crc kubenswrapper[4979]: I1204 12:02:28.893697 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:02:28 crc kubenswrapper[4979]: I1204 12:02:28.914629 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7hhk\" (UniqueName: \"kubernetes.io/projected/dd8830de-44cb-4015-98f2-2d55ffdccf52-kube-api-access-d7hhk\") pod \"kube-state-metrics-0\" (UID: \"dd8830de-44cb-4015-98f2-2d55ffdccf52\") " pod="openstack/kube-state-metrics-0" Dec 04 12:02:29 crc kubenswrapper[4979]: I1204 12:02:29.015933 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7hhk\" (UniqueName: \"kubernetes.io/projected/dd8830de-44cb-4015-98f2-2d55ffdccf52-kube-api-access-d7hhk\") pod \"kube-state-metrics-0\" (UID: \"dd8830de-44cb-4015-98f2-2d55ffdccf52\") " pod="openstack/kube-state-metrics-0" Dec 04 12:02:29 crc kubenswrapper[4979]: I1204 12:02:29.036418 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7hhk\" (UniqueName: \"kubernetes.io/projected/dd8830de-44cb-4015-98f2-2d55ffdccf52-kube-api-access-d7hhk\") pod \"kube-state-metrics-0\" (UID: \"dd8830de-44cb-4015-98f2-2d55ffdccf52\") " pod="openstack/kube-state-metrics-0" Dec 04 12:02:29 crc kubenswrapper[4979]: I1204 12:02:29.229503 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.284054 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x2hnf"] Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.285493 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.291284 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-x54hj"] Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.293617 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.295008 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-sf4kw" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.295045 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.295337 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.328505 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x2hnf"] Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.337459 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-x54hj"] Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368432 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-log-ovn\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368513 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-run\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368550 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-ovn-controller-tls-certs\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368574 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-log\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368626 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-lib\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368708 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdps4\" (UniqueName: \"kubernetes.io/projected/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-kube-api-access-vdps4\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368728 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09e1a1f2-c92b-46be-835a-8ca671f26472-scripts\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368780 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run-ovn\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368803 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98qtc\" (UniqueName: \"kubernetes.io/projected/09e1a1f2-c92b-46be-835a-8ca671f26472-kube-api-access-98qtc\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368821 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368878 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-etc-ovs\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368906 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-scripts\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.368927 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-combined-ca-bundle\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.469743 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-run\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.469809 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-ovn-controller-tls-certs\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.469837 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-log\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.469888 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-lib\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470197 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdps4\" (UniqueName: \"kubernetes.io/projected/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-kube-api-access-vdps4\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470352 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-run\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470458 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-lib\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470545 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-log\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470594 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09e1a1f2-c92b-46be-835a-8ca671f26472-scripts\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470630 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run-ovn\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470666 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qtc\" (UniqueName: \"kubernetes.io/projected/09e1a1f2-c92b-46be-835a-8ca671f26472-kube-api-access-98qtc\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470694 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470723 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-etc-ovs\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470762 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-scripts\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470788 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-combined-ca-bundle\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.470837 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-log-ovn\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.471008 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-log-ovn\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.471108 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run-ovn\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.471420 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.471574 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-etc-ovs\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.472628 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09e1a1f2-c92b-46be-835a-8ca671f26472-scripts\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.475534 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-scripts\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.479618 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-ovn-controller-tls-certs\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.483237 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-combined-ca-bundle\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.487723 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdps4\" (UniqueName: \"kubernetes.io/projected/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-kube-api-access-vdps4\") pod \"ovn-controller-x2hnf\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.492419 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qtc\" (UniqueName: \"kubernetes.io/projected/09e1a1f2-c92b-46be-835a-8ca671f26472-kube-api-access-98qtc\") pod \"ovn-controller-ovs-x54hj\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.513680 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f763e763-10df-408d-9452-84634d47b455","Type":"ContainerStarted","Data":"298d0f7088901bd1bf19e095b09802198329ed8cbbe172cad02e5d04b8488026"} Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.649615 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:32 crc kubenswrapper[4979]: I1204 12:02:32.657633 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.162981 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.164154 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.166073 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.166135 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.166256 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.166430 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-k9tnt" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.173208 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.173367 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181705 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181758 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181809 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkp69\" (UniqueName: \"kubernetes.io/projected/ef961b3d-b927-42e5-91b9-eb966bdee861-kube-api-access-gkp69\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181862 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181900 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181939 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.181973 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.182010 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-config\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283240 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-config\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283369 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283401 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283425 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkp69\" (UniqueName: \"kubernetes.io/projected/ef961b3d-b927-42e5-91b9-eb966bdee861-kube-api-access-gkp69\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283454 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283483 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.283846 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.284610 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-config\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.285578 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.285673 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.286009 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.286906 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.289937 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.291259 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.297103 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.301063 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkp69\" (UniqueName: \"kubernetes.io/projected/ef961b3d-b927-42e5-91b9-eb966bdee861-kube-api-access-gkp69\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.312582 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:33 crc kubenswrapper[4979]: I1204 12:02:33.482650 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.905719 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.907534 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.909411 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.911086 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.911398 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-nr2bj" Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.911508 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 12:02:36 crc kubenswrapper[4979]: I1204 12:02:36.933642 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.046867 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047121 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047211 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-config\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047381 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047517 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047662 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qrt2\" (UniqueName: \"kubernetes.io/projected/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-kube-api-access-8qrt2\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047807 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.047900 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149587 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149663 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149693 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qrt2\" (UniqueName: \"kubernetes.io/projected/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-kube-api-access-8qrt2\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149754 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149786 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149840 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149871 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.149904 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-config\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.150235 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.150456 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.151083 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-config\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.151115 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.156632 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.157629 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.157637 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.188060 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qrt2\" (UniqueName: \"kubernetes.io/projected/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-kube-api-access-8qrt2\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.189376 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:37 crc kubenswrapper[4979]: I1204 12:02:37.230327 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 12:02:38 crc kubenswrapper[4979]: I1204 12:02:38.558586 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3791d23-03e1-4c8d-b44e-738eff9222d0","Type":"ContainerStarted","Data":"0e7aa55bc3b8f0c516ee6b7c961e4d6cdb3d45ae3309ae2fb4bb62fad283ab6b"} Dec 04 12:02:47 crc kubenswrapper[4979]: E1204 12:02:47.996145 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 04 12:02:47 crc kubenswrapper[4979]: E1204 12:02:47.997072 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jkr8v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(0fad7195-d19a-48ce-ad3d-4c67b53d2974): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:02:47 crc kubenswrapper[4979]: E1204 12:02:47.999054 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" Dec 04 12:02:48 crc kubenswrapper[4979]: E1204 12:02:48.652790 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.112417 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.112834 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vx6s2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-rm27j_openstack(52fdc353-c66e-46d4-9021-869535523f3d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.113896 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" podUID="52fdc353-c66e-46d4-9021-869535523f3d" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.119985 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.120167 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqqh8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-p4zbn_openstack(3c1e3480-dc4a-438c-a6d0-ede582da1a7e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.121511 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.140760 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.140936 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jzm4n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(0e592d2c-bb50-44b9-bd85-fdf94ee1ac48): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.142493 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.142540 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.142568 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6dv8f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-jj8xn_openstack(d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.143722 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" podUID="d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.217861 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.218001 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5zn6b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-kmnvq_openstack(e9a2aa4b-cea0-40e4-94b8-d19edadde1e7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.219182 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" podUID="e9a2aa4b-cea0-40e4-94b8-d19edadde1e7" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.678375 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.678375 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" Dec 04 12:02:52 crc kubenswrapper[4979]: E1204 12:02:52.678921 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" podUID="e9a2aa4b-cea0-40e4-94b8-d19edadde1e7" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.112351 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.224884 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-dns-svc\") pod \"52fdc353-c66e-46d4-9021-869535523f3d\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.225328 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-config\") pod \"52fdc353-c66e-46d4-9021-869535523f3d\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.225385 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "52fdc353-c66e-46d4-9021-869535523f3d" (UID: "52fdc353-c66e-46d4-9021-869535523f3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.225438 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx6s2\" (UniqueName: \"kubernetes.io/projected/52fdc353-c66e-46d4-9021-869535523f3d-kube-api-access-vx6s2\") pod \"52fdc353-c66e-46d4-9021-869535523f3d\" (UID: \"52fdc353-c66e-46d4-9021-869535523f3d\") " Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.225766 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.225762 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-config" (OuterVolumeSpecName: "config") pod "52fdc353-c66e-46d4-9021-869535523f3d" (UID: "52fdc353-c66e-46d4-9021-869535523f3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.226564 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.244553 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/52fdc353-c66e-46d4-9021-869535523f3d-kube-api-access-vx6s2" (OuterVolumeSpecName: "kube-api-access-vx6s2") pod "52fdc353-c66e-46d4-9021-869535523f3d" (UID: "52fdc353-c66e-46d4-9021-869535523f3d"). InnerVolumeSpecName "kube-api-access-vx6s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.316836 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.326282 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-config\") pod \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.326367 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dv8f\" (UniqueName: \"kubernetes.io/projected/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-kube-api-access-6dv8f\") pod \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\" (UID: \"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc\") " Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.326813 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/52fdc353-c66e-46d4-9021-869535523f3d-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.326829 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx6s2\" (UniqueName: \"kubernetes.io/projected/52fdc353-c66e-46d4-9021-869535523f3d-kube-api-access-vx6s2\") on node \"crc\" DevicePath \"\"" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.327543 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-config" (OuterVolumeSpecName: "config") pod "d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc" (UID: "d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.335003 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-kube-api-access-6dv8f" (OuterVolumeSpecName: "kube-api-access-6dv8f") pod "d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc" (UID: "d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc"). InnerVolumeSpecName "kube-api-access-6dv8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.429341 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.429407 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dv8f\" (UniqueName: \"kubernetes.io/projected/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc-kube-api-access-6dv8f\") on node \"crc\" DevicePath \"\"" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.508117 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x2hnf"] Dec 04 12:02:53 crc kubenswrapper[4979]: W1204 12:02:53.514000 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2853b847_4bf0_4d0d_af7b_ca8bc87a9c99.slice/crio-a0d3997c62f68fc25228cb0f599a184a35be015cfd2038e9541835c1f07bb64f WatchSource:0}: Error finding container a0d3997c62f68fc25228cb0f599a184a35be015cfd2038e9541835c1f07bb64f: Status 404 returned error can't find the container with id a0d3997c62f68fc25228cb0f599a184a35be015cfd2038e9541835c1f07bb64f Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.697694 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7130263f-15db-4a87-a078-9c346739eced","Type":"ContainerStarted","Data":"ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8"} Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.710667 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.718503 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f763e763-10df-408d-9452-84634d47b455","Type":"ContainerStarted","Data":"b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c"} Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.719005 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.721952 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" event={"ID":"52fdc353-c66e-46d4-9021-869535523f3d","Type":"ContainerDied","Data":"22a815c1dec64065e00120719ea5be2eabba2e7e5b64735c1259efdf42df8369"} Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.722075 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-rm27j" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.735015 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3791d23-03e1-4c8d-b44e-738eff9222d0","Type":"ContainerStarted","Data":"7c15680c11567bfbc6297277c433787fb22a46e130fb931cdb93732c47665bb6"} Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.742979 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"dd8830de-44cb-4015-98f2-2d55ffdccf52","Type":"ContainerStarted","Data":"d0e20d5ba03db06272c4946472538fa78cca5f30fc71f6b0f192583fd5958252"} Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.745542 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.745546 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-jj8xn" event={"ID":"d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc","Type":"ContainerDied","Data":"dfe83af52541f0bfaae7fb5b9fe900bc443f1041742fa6c5616250c9e71b9ac1"} Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.747051 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf" event={"ID":"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99","Type":"ContainerStarted","Data":"a0d3997c62f68fc25228cb0f599a184a35be015cfd2038e9541835c1f07bb64f"} Dec 04 12:02:53 crc kubenswrapper[4979]: W1204 12:02:53.787963 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09e1a1f2_c92b_46be_835a_8ca671f26472.slice/crio-15dbebed4170dd959a37049f55a4bf615ac2122ee366747d46f50e111773915c WatchSource:0}: Error finding container 15dbebed4170dd959a37049f55a4bf615ac2122ee366747d46f50e111773915c: Status 404 returned error can't find the container with id 15dbebed4170dd959a37049f55a4bf615ac2122ee366747d46f50e111773915c Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.799344 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-x54hj"] Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.800862 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=6.237076083 podStartE2EDuration="26.800843021s" podCreationTimestamp="2025-12-04 12:02:27 +0000 UTC" firstStartedPulling="2025-12-04 12:02:32.315559855 +0000 UTC m=+1176.589855659" lastFinishedPulling="2025-12-04 12:02:52.879326683 +0000 UTC m=+1197.153622597" observedRunningTime="2025-12-04 12:02:53.792873054 +0000 UTC m=+1198.067168858" watchObservedRunningTime="2025-12-04 12:02:53.800843021 +0000 UTC m=+1198.075138825" Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.831180 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rm27j"] Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.841751 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-rm27j"] Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.874419 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jj8xn"] Dec 04 12:02:53 crc kubenswrapper[4979]: I1204 12:02:53.879019 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-jj8xn"] Dec 04 12:02:54 crc kubenswrapper[4979]: I1204 12:02:54.209131 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="52fdc353-c66e-46d4-9021-869535523f3d" path="/var/lib/kubelet/pods/52fdc353-c66e-46d4-9021-869535523f3d/volumes" Dec 04 12:02:54 crc kubenswrapper[4979]: I1204 12:02:54.209590 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc" path="/var/lib/kubelet/pods/d6118fae-fd8f-4fb3-8b7d-d4768a4b84bc/volumes" Dec 04 12:02:54 crc kubenswrapper[4979]: I1204 12:02:54.571628 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 12:02:54 crc kubenswrapper[4979]: W1204 12:02:54.761931 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbdd8e97d_9bc0_4e10_984b_7d92698e8bd6.slice/crio-566bb3521fcc918076339895b00651591a9d329423b5a31229a6b38c106e9c2d WatchSource:0}: Error finding container 566bb3521fcc918076339895b00651591a9d329423b5a31229a6b38c106e9c2d: Status 404 returned error can't find the container with id 566bb3521fcc918076339895b00651591a9d329423b5a31229a6b38c106e9c2d Dec 04 12:02:54 crc kubenswrapper[4979]: I1204 12:02:54.763819 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ef961b3d-b927-42e5-91b9-eb966bdee861","Type":"ContainerStarted","Data":"2f24c9ff025d883e5b4288cd6adff9b63f9decd968e600453fdad6a9eb5c359b"} Dec 04 12:02:54 crc kubenswrapper[4979]: I1204 12:02:54.768501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerStarted","Data":"15dbebed4170dd959a37049f55a4bf615ac2122ee366747d46f50e111773915c"} Dec 04 12:02:55 crc kubenswrapper[4979]: I1204 12:02:55.776700 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6","Type":"ContainerStarted","Data":"566bb3521fcc918076339895b00651591a9d329423b5a31229a6b38c106e9c2d"} Dec 04 12:02:56 crc kubenswrapper[4979]: I1204 12:02:56.787563 4979 generic.go:334] "Generic (PLEG): container finished" podID="7130263f-15db-4a87-a078-9c346739eced" containerID="ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8" exitCode=0 Dec 04 12:02:56 crc kubenswrapper[4979]: I1204 12:02:56.787673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7130263f-15db-4a87-a078-9c346739eced","Type":"ContainerDied","Data":"ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8"} Dec 04 12:02:56 crc kubenswrapper[4979]: I1204 12:02:56.790770 4979 generic.go:334] "Generic (PLEG): container finished" podID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerID="7c15680c11567bfbc6297277c433787fb22a46e130fb931cdb93732c47665bb6" exitCode=0 Dec 04 12:02:56 crc kubenswrapper[4979]: I1204 12:02:56.790821 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3791d23-03e1-4c8d-b44e-738eff9222d0","Type":"ContainerDied","Data":"7c15680c11567bfbc6297277c433787fb22a46e130fb931cdb93732c47665bb6"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.833015 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf" event={"ID":"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99","Type":"ContainerStarted","Data":"5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.833603 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-x2hnf" Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.840903 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6","Type":"ContainerStarted","Data":"0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.851589 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7130263f-15db-4a87-a078-9c346739eced","Type":"ContainerStarted","Data":"1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.852013 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-x2hnf" podStartSLOduration=22.823072424 podStartE2EDuration="27.852002961s" podCreationTimestamp="2025-12-04 12:02:32 +0000 UTC" firstStartedPulling="2025-12-04 12:02:53.516982108 +0000 UTC m=+1197.791277912" lastFinishedPulling="2025-12-04 12:02:58.545912645 +0000 UTC m=+1202.820208449" observedRunningTime="2025-12-04 12:02:59.849625526 +0000 UTC m=+1204.123921350" watchObservedRunningTime="2025-12-04 12:02:59.852002961 +0000 UTC m=+1204.126298765" Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.855919 4979 generic.go:334] "Generic (PLEG): container finished" podID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerID="42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a" exitCode=0 Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.855950 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerDied","Data":"42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.858403 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ef961b3d-b927-42e5-91b9-eb966bdee861","Type":"ContainerStarted","Data":"c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.861404 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3791d23-03e1-4c8d-b44e-738eff9222d0","Type":"ContainerStarted","Data":"fe043f249efb9dfd8bb0e3a20f1360262cf5b2c026e1870681a1fabdae2c45d6"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.865082 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"dd8830de-44cb-4015-98f2-2d55ffdccf52","Type":"ContainerStarted","Data":"54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036"} Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.865244 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.871433 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=9.584325577 podStartE2EDuration="35.869115146s" podCreationTimestamp="2025-12-04 12:02:24 +0000 UTC" firstStartedPulling="2025-12-04 12:02:26.501242739 +0000 UTC m=+1170.775538543" lastFinishedPulling="2025-12-04 12:02:52.786032308 +0000 UTC m=+1197.060328112" observedRunningTime="2025-12-04 12:02:59.869093395 +0000 UTC m=+1204.143389209" watchObservedRunningTime="2025-12-04 12:02:59.869115146 +0000 UTC m=+1204.143410950" Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.935323 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=26.682511295 podStartE2EDuration="31.935283633s" podCreationTimestamp="2025-12-04 12:02:28 +0000 UTC" firstStartedPulling="2025-12-04 12:02:53.323362808 +0000 UTC m=+1197.597658622" lastFinishedPulling="2025-12-04 12:02:58.576135156 +0000 UTC m=+1202.850430960" observedRunningTime="2025-12-04 12:02:59.909579555 +0000 UTC m=+1204.183875359" watchObservedRunningTime="2025-12-04 12:02:59.935283633 +0000 UTC m=+1204.209579437" Dec 04 12:02:59 crc kubenswrapper[4979]: I1204 12:02:59.937262 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=19.66758825 podStartE2EDuration="34.937253398s" podCreationTimestamp="2025-12-04 12:02:25 +0000 UTC" firstStartedPulling="2025-12-04 12:02:37.61279643 +0000 UTC m=+1181.887092234" lastFinishedPulling="2025-12-04 12:02:52.882461578 +0000 UTC m=+1197.156757382" observedRunningTime="2025-12-04 12:02:59.928607243 +0000 UTC m=+1204.202903047" watchObservedRunningTime="2025-12-04 12:02:59.937253398 +0000 UTC m=+1204.211549202" Dec 04 12:03:00 crc kubenswrapper[4979]: I1204 12:03:00.876386 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerStarted","Data":"1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e"} Dec 04 12:03:00 crc kubenswrapper[4979]: I1204 12:03:00.876987 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerStarted","Data":"64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c"} Dec 04 12:03:00 crc kubenswrapper[4979]: I1204 12:03:00.877037 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:03:00 crc kubenswrapper[4979]: I1204 12:03:00.900466 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-x54hj" podStartSLOduration=24.155323352 podStartE2EDuration="28.900446848s" podCreationTimestamp="2025-12-04 12:02:32 +0000 UTC" firstStartedPulling="2025-12-04 12:02:53.79051479 +0000 UTC m=+1198.064810594" lastFinishedPulling="2025-12-04 12:02:58.535638286 +0000 UTC m=+1202.809934090" observedRunningTime="2025-12-04 12:03:00.895630776 +0000 UTC m=+1205.169926590" watchObservedRunningTime="2025-12-04 12:03:00.900446848 +0000 UTC m=+1205.174742662" Dec 04 12:03:01 crc kubenswrapper[4979]: I1204 12:03:01.895482 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:03:02 crc kubenswrapper[4979]: I1204 12:03:02.642998 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 12:03:02 crc kubenswrapper[4979]: I1204 12:03:02.902896 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6","Type":"ContainerStarted","Data":"da2072151fb1ba04ce685c63911129fe34aff9fc376cba5805e8cd1d4a2e02b6"} Dec 04 12:03:02 crc kubenswrapper[4979]: I1204 12:03:02.905401 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ef961b3d-b927-42e5-91b9-eb966bdee861","Type":"ContainerStarted","Data":"43237a8cc1d10bdde09a5a677588338264c1215b57330e0176e9bd4306ed5bf6"} Dec 04 12:03:02 crc kubenswrapper[4979]: I1204 12:03:02.926976 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=20.791221018999998 podStartE2EDuration="27.926959558s" podCreationTimestamp="2025-12-04 12:02:35 +0000 UTC" firstStartedPulling="2025-12-04 12:02:54.768915963 +0000 UTC m=+1199.043211777" lastFinishedPulling="2025-12-04 12:03:01.904654502 +0000 UTC m=+1206.178950316" observedRunningTime="2025-12-04 12:03:02.923897705 +0000 UTC m=+1207.198193509" watchObservedRunningTime="2025-12-04 12:03:02.926959558 +0000 UTC m=+1207.201255362" Dec 04 12:03:02 crc kubenswrapper[4979]: I1204 12:03:02.954429 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=22.806033943 podStartE2EDuration="30.954408944s" podCreationTimestamp="2025-12-04 12:02:32 +0000 UTC" firstStartedPulling="2025-12-04 12:02:53.721962108 +0000 UTC m=+1197.996257912" lastFinishedPulling="2025-12-04 12:03:01.870337089 +0000 UTC m=+1206.144632913" observedRunningTime="2025-12-04 12:03:02.948887344 +0000 UTC m=+1207.223183138" watchObservedRunningTime="2025-12-04 12:03:02.954408944 +0000 UTC m=+1207.228704748" Dec 04 12:03:03 crc kubenswrapper[4979]: I1204 12:03:03.483814 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 12:03:03 crc kubenswrapper[4979]: I1204 12:03:03.483883 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 12:03:03 crc kubenswrapper[4979]: I1204 12:03:03.537522 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 12:03:03 crc kubenswrapper[4979]: I1204 12:03:03.912610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0fad7195-d19a-48ce-ad3d-4c67b53d2974","Type":"ContainerStarted","Data":"ddc7fe209b64e40b2ff31581c3560cda53976230e0b9d95e54824b9db34296d8"} Dec 04 12:03:03 crc kubenswrapper[4979]: I1204 12:03:03.960596 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.231010 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.270077 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kmnvq"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.304639 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-prxtm"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.305876 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.307070 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.309128 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.317335 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-prxtm"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.326838 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-v6652"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.327844 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.340855 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.352282 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v6652"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.419922 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.419964 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.420004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-config\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.420080 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wt6c\" (UniqueName: \"kubernetes.io/projected/37734cd3-9b2b-47e0-85bb-3161d95229d3-kube-api-access-4wt6c\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.521538 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovs-rundir\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.521900 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4nnm\" (UniqueName: \"kubernetes.io/projected/70791582-cb57-4564-8a5b-9953f4f22c23-kube-api-access-f4nnm\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522019 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wt6c\" (UniqueName: \"kubernetes.io/projected/37734cd3-9b2b-47e0-85bb-3161d95229d3-kube-api-access-4wt6c\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522095 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522119 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522143 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-combined-ca-bundle\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522193 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522318 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-config\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522415 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovn-rundir\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522476 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70791582-cb57-4564-8a5b-9953f4f22c23-config\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.522959 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-dns-svc\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.523286 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-config\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.524522 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-ovsdbserver-nb\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.552940 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wt6c\" (UniqueName: \"kubernetes.io/projected/37734cd3-9b2b-47e0-85bb-3161d95229d3-kube-api-access-4wt6c\") pod \"dnsmasq-dns-7fd796d7df-prxtm\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.623966 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-combined-ca-bundle\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624014 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624055 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovn-rundir\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624077 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70791582-cb57-4564-8a5b-9953f4f22c23-config\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624128 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovs-rundir\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624153 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4nnm\" (UniqueName: \"kubernetes.io/projected/70791582-cb57-4564-8a5b-9953f4f22c23-kube-api-access-f4nnm\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624510 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovn-rundir\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624577 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovs-rundir\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.624949 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70791582-cb57-4564-8a5b-9953f4f22c23-config\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.626288 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.629055 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-combined-ca-bundle\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.633844 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.650353 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p4zbn"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.658949 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4nnm\" (UniqueName: \"kubernetes.io/projected/70791582-cb57-4564-8a5b-9953f4f22c23-kube-api-access-f4nnm\") pod \"ovn-controller-metrics-v6652\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.677612 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.685455 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kpn2l"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.686983 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.691536 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.699278 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kpn2l"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.797604 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.827511 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.827585 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.827636 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5qvz\" (UniqueName: \"kubernetes.io/projected/a6666948-dde4-490a-93af-d5494b1e8243-kube-api-access-p5qvz\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.827702 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.827725 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-config\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.919877 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" event={"ID":"3c1e3480-dc4a-438c-a6d0-ede582da1a7e","Type":"ContainerStarted","Data":"2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f"} Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.920053 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" containerName="init" containerID="cri-o://2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f" gracePeriod=10 Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.923922 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" event={"ID":"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7","Type":"ContainerDied","Data":"b183cdfe9e048ae2c7fc460930178e0df568919de7c2a9406d0ea18085f61373"} Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.924050 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kmnvq" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.925249 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.931603 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-dns-svc\") pod \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.931801 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-config\") pod \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.932017 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zn6b\" (UniqueName: \"kubernetes.io/projected/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-kube-api-access-5zn6b\") pod \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\" (UID: \"e9a2aa4b-cea0-40e4-94b8-d19edadde1e7\") " Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.932310 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9a2aa4b-cea0-40e4-94b8-d19edadde1e7" (UID: "e9a2aa4b-cea0-40e4-94b8-d19edadde1e7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.932413 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-config" (OuterVolumeSpecName: "config") pod "e9a2aa4b-cea0-40e4-94b8-d19edadde1e7" (UID: "e9a2aa4b-cea0-40e4-94b8-d19edadde1e7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.932448 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.932535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.932758 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5qvz\" (UniqueName: \"kubernetes.io/projected/a6666948-dde4-490a-93af-d5494b1e8243-kube-api-access-p5qvz\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.933058 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.933157 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.933173 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-config\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.933394 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.933419 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.935153 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.936544 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.946865 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-config\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.952711 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-kube-api-access-5zn6b" (OuterVolumeSpecName: "kube-api-access-5zn6b") pod "e9a2aa4b-cea0-40e4-94b8-d19edadde1e7" (UID: "e9a2aa4b-cea0-40e4-94b8-d19edadde1e7"). InnerVolumeSpecName "kube-api-access-5zn6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.954632 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5qvz\" (UniqueName: \"kubernetes.io/projected/a6666948-dde4-490a-93af-d5494b1e8243-kube-api-access-p5qvz\") pod \"dnsmasq-dns-86db49b7ff-kpn2l\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.995391 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-v6652"] Dec 04 12:03:04 crc kubenswrapper[4979]: I1204 12:03:04.996441 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.035579 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zn6b\" (UniqueName: \"kubernetes.io/projected/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7-kube-api-access-5zn6b\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.121945 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.157781 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.159041 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.164232 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.164519 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.165851 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.166006 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-9jsj4" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.167136 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.172524 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-prxtm"] Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.243938 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.244328 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-scripts\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.244371 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.244394 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.244415 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.244482 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-config\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.244504 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrtz8\" (UniqueName: \"kubernetes.io/projected/d5e30fc1-acc8-448a-a9e8-490bda22e24e-kube-api-access-zrtz8\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.247799 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.330359 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kmnvq"] Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.337367 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kmnvq"] Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.348461 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-dns-svc\") pod \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.348615 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-config\") pod \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.348705 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqqh8\" (UniqueName: \"kubernetes.io/projected/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-kube-api-access-nqqh8\") pod \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\" (UID: \"3c1e3480-dc4a-438c-a6d0-ede582da1a7e\") " Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.348892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.348923 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.348943 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.349036 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-config\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.349064 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrtz8\" (UniqueName: \"kubernetes.io/projected/d5e30fc1-acc8-448a-a9e8-490bda22e24e-kube-api-access-zrtz8\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.349135 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.349167 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-scripts\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.350000 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-scripts\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.350753 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.359711 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-config\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.367976 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.397208 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.399254 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-kube-api-access-nqqh8" (OuterVolumeSpecName: "kube-api-access-nqqh8") pod "3c1e3480-dc4a-438c-a6d0-ede582da1a7e" (UID: "3c1e3480-dc4a-438c-a6d0-ede582da1a7e"). InnerVolumeSpecName "kube-api-access-nqqh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.424281 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.424936 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrtz8\" (UniqueName: \"kubernetes.io/projected/d5e30fc1-acc8-448a-a9e8-490bda22e24e-kube-api-access-zrtz8\") pod \"ovn-northd-0\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.455523 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqqh8\" (UniqueName: \"kubernetes.io/projected/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-kube-api-access-nqqh8\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.460143 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3c1e3480-dc4a-438c-a6d0-ede582da1a7e" (UID: "3c1e3480-dc4a-438c-a6d0-ede582da1a7e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.465834 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-config" (OuterVolumeSpecName: "config") pod "3c1e3480-dc4a-438c-a6d0-ede582da1a7e" (UID: "3c1e3480-dc4a-438c-a6d0-ede582da1a7e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.505703 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.558018 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.558044 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3c1e3480-dc4a-438c-a6d0-ede582da1a7e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:05 crc kubenswrapper[4979]: W1204 12:03:05.643984 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6666948_dde4_490a_93af_d5494b1e8243.slice/crio-5d22631616b602de8536d56ef86994da40474e7622c6049f34e2d2d4a8a21fc1 WatchSource:0}: Error finding container 5d22631616b602de8536d56ef86994da40474e7622c6049f34e2d2d4a8a21fc1: Status 404 returned error can't find the container with id 5d22631616b602de8536d56ef86994da40474e7622c6049f34e2d2d4a8a21fc1 Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.658627 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kpn2l"] Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.809893 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.810536 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.891489 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.946006 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v6652" event={"ID":"70791582-cb57-4564-8a5b-9953f4f22c23","Type":"ContainerStarted","Data":"992bc26c538b977fe854be958e44ced58bdcb0d9b0705720d5e5a3f11286c1a4"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.946176 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v6652" event={"ID":"70791582-cb57-4564-8a5b-9953f4f22c23","Type":"ContainerStarted","Data":"02d87c993f9049eaa0cb96903e59776aaa45e7d153e10fab1bbf6bb9aeef4ff1"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.961263 4979 generic.go:334] "Generic (PLEG): container finished" podID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerID="8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51" exitCode=0 Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.961319 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" event={"ID":"37734cd3-9b2b-47e0-85bb-3161d95229d3","Type":"ContainerDied","Data":"8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.961360 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" event={"ID":"37734cd3-9b2b-47e0-85bb-3161d95229d3","Type":"ContainerStarted","Data":"8e721dc08e1b1e75e17d9779c958399c37da861bcb668f081e7da2ba01c290b6"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.965687 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6666948-dde4-490a-93af-d5494b1e8243" containerID="cc35a467283186803a3147f7ae91b7b31c57a2058c85bd55b0f75ef4ba36ebf2" exitCode=0 Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.965787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" event={"ID":"a6666948-dde4-490a-93af-d5494b1e8243","Type":"ContainerDied","Data":"cc35a467283186803a3147f7ae91b7b31c57a2058c85bd55b0f75ef4ba36ebf2"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.965814 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" event={"ID":"a6666948-dde4-490a-93af-d5494b1e8243","Type":"ContainerStarted","Data":"5d22631616b602de8536d56ef86994da40474e7622c6049f34e2d2d4a8a21fc1"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.966921 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-v6652" podStartSLOduration=1.9669036530000001 podStartE2EDuration="1.966903653s" podCreationTimestamp="2025-12-04 12:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:05.963672956 +0000 UTC m=+1210.237968770" watchObservedRunningTime="2025-12-04 12:03:05.966903653 +0000 UTC m=+1210.241199467" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.968696 4979 generic.go:334] "Generic (PLEG): container finished" podID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" containerID="2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f" exitCode=0 Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.968739 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" event={"ID":"3c1e3480-dc4a-438c-a6d0-ede582da1a7e","Type":"ContainerDied","Data":"2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.968777 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" event={"ID":"3c1e3480-dc4a-438c-a6d0-ede582da1a7e","Type":"ContainerDied","Data":"e7c59f7abed59d9f5ff4c6889e2d2cc265d46fbdb0a7be490c19f85650c8ee01"} Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.968799 4979 scope.go:117] "RemoveContainer" containerID="2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.968920 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-p4zbn" Dec 04 12:03:05 crc kubenswrapper[4979]: I1204 12:03:05.993714 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.077289 4979 scope.go:117] "RemoveContainer" containerID="2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f" Dec 04 12:03:06 crc kubenswrapper[4979]: E1204 12:03:06.089375 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f\": container with ID starting with 2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f not found: ID does not exist" containerID="2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f" Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.089429 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f"} err="failed to get container status \"2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f\": rpc error: code = NotFound desc = could not find container \"2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f\": container with ID starting with 2b9dac7256a533d862a8c3d56cf99bca8774709c8e3328d110599b47868aca0f not found: ID does not exist" Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.094906 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p4zbn"] Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.108237 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.117927 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-p4zbn"] Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.214718 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" path="/var/lib/kubelet/pods/3c1e3480-dc4a-438c-a6d0-ede582da1a7e/volumes" Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.215721 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a2aa4b-cea0-40e4-94b8-d19edadde1e7" path="/var/lib/kubelet/pods/e9a2aa4b-cea0-40e4-94b8-d19edadde1e7/volumes" Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.979265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d5e30fc1-acc8-448a-a9e8-490bda22e24e","Type":"ContainerStarted","Data":"ba4881247e039da9ed32cbe3e313560113fd6c9ec118429556cf7aecb95172e6"} Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.981563 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" event={"ID":"a6666948-dde4-490a-93af-d5494b1e8243","Type":"ContainerStarted","Data":"f1c694f83174118748120db5cc618d6289efeb90b3ddca198eb54de151f03ee5"} Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.982852 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:06 crc kubenswrapper[4979]: I1204 12:03:06.988355 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" event={"ID":"37734cd3-9b2b-47e0-85bb-3161d95229d3","Type":"ContainerStarted","Data":"6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041"} Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.027868 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" podStartSLOduration=3.027845939 podStartE2EDuration="3.027845939s" podCreationTimestamp="2025-12-04 12:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:07.008809242 +0000 UTC m=+1211.283105046" watchObservedRunningTime="2025-12-04 12:03:07.027845939 +0000 UTC m=+1211.302141743" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.033531 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" podStartSLOduration=3.033516683 podStartE2EDuration="3.033516683s" podCreationTimestamp="2025-12-04 12:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:07.024763656 +0000 UTC m=+1211.299059460" watchObservedRunningTime="2025-12-04 12:03:07.033516683 +0000 UTC m=+1211.307812487" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.245330 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ac67-account-create-update-ztshn"] Dec 04 12:03:07 crc kubenswrapper[4979]: E1204 12:03:07.245732 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" containerName="init" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.245747 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" containerName="init" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.245900 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c1e3480-dc4a-438c-a6d0-ede582da1a7e" containerName="init" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.246456 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.250549 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.253402 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ac67-account-create-update-ztshn"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.266957 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.270442 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.288838 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-zxkmj"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.288959 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09068fde-06eb-4075-b4aa-78c8de781ced-operator-scripts\") pod \"keystone-ac67-account-create-update-ztshn\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.289038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m7nh\" (UniqueName: \"kubernetes.io/projected/09068fde-06eb-4075-b4aa-78c8de781ced-kube-api-access-7m7nh\") pod \"keystone-ac67-account-create-update-ztshn\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.290137 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.338998 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zxkmj"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.390404 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m7nh\" (UniqueName: \"kubernetes.io/projected/09068fde-06eb-4075-b4aa-78c8de781ced-kube-api-access-7m7nh\") pod \"keystone-ac67-account-create-update-ztshn\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.393119 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e82e89-f20a-451f-b992-5b07d77b1e14-operator-scripts\") pod \"keystone-db-create-zxkmj\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.393186 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ptk6\" (UniqueName: \"kubernetes.io/projected/35e82e89-f20a-451f-b992-5b07d77b1e14-kube-api-access-2ptk6\") pod \"keystone-db-create-zxkmj\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.393217 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09068fde-06eb-4075-b4aa-78c8de781ced-operator-scripts\") pod \"keystone-ac67-account-create-update-ztshn\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.394032 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09068fde-06eb-4075-b4aa-78c8de781ced-operator-scripts\") pod \"keystone-ac67-account-create-update-ztshn\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.395867 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.420070 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m7nh\" (UniqueName: \"kubernetes.io/projected/09068fde-06eb-4075-b4aa-78c8de781ced-kube-api-access-7m7nh\") pod \"keystone-ac67-account-create-update-ztshn\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.494424 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e82e89-f20a-451f-b992-5b07d77b1e14-operator-scripts\") pod \"keystone-db-create-zxkmj\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.494467 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ptk6\" (UniqueName: \"kubernetes.io/projected/35e82e89-f20a-451f-b992-5b07d77b1e14-kube-api-access-2ptk6\") pod \"keystone-db-create-zxkmj\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.495336 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e82e89-f20a-451f-b992-5b07d77b1e14-operator-scripts\") pod \"keystone-db-create-zxkmj\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.510690 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ptk6\" (UniqueName: \"kubernetes.io/projected/35e82e89-f20a-451f-b992-5b07d77b1e14-kube-api-access-2ptk6\") pod \"keystone-db-create-zxkmj\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.574838 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.603522 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-wl8cj"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.604518 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.609217 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.621503 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wl8cj"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.696927 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-operator-scripts\") pod \"placement-db-create-wl8cj\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.697016 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrn28\" (UniqueName: \"kubernetes.io/projected/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-kube-api-access-wrn28\") pod \"placement-db-create-wl8cj\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.722182 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-375f-account-create-update-bh966"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.723321 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.725564 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.745891 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-375f-account-create-update-bh966"] Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.800715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n2xt\" (UniqueName: \"kubernetes.io/projected/91e3fea1-7f8b-454c-9033-75cd30247761-kube-api-access-4n2xt\") pod \"placement-375f-account-create-update-bh966\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.800756 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-operator-scripts\") pod \"placement-db-create-wl8cj\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.800804 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrn28\" (UniqueName: \"kubernetes.io/projected/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-kube-api-access-wrn28\") pod \"placement-db-create-wl8cj\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.800870 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91e3fea1-7f8b-454c-9033-75cd30247761-operator-scripts\") pod \"placement-375f-account-create-update-bh966\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.801421 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-operator-scripts\") pod \"placement-db-create-wl8cj\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.821443 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrn28\" (UniqueName: \"kubernetes.io/projected/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-kube-api-access-wrn28\") pod \"placement-db-create-wl8cj\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.904094 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n2xt\" (UniqueName: \"kubernetes.io/projected/91e3fea1-7f8b-454c-9033-75cd30247761-kube-api-access-4n2xt\") pod \"placement-375f-account-create-update-bh966\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.905170 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91e3fea1-7f8b-454c-9033-75cd30247761-operator-scripts\") pod \"placement-375f-account-create-update-bh966\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.905932 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91e3fea1-7f8b-454c-9033-75cd30247761-operator-scripts\") pod \"placement-375f-account-create-update-bh966\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.926793 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n2xt\" (UniqueName: \"kubernetes.io/projected/91e3fea1-7f8b-454c-9033-75cd30247761-kube-api-access-4n2xt\") pod \"placement-375f-account-create-update-bh966\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.997054 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d5e30fc1-acc8-448a-a9e8-490bda22e24e","Type":"ContainerStarted","Data":"104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024"} Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.997104 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d5e30fc1-acc8-448a-a9e8-490bda22e24e","Type":"ContainerStarted","Data":"4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a"} Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.997824 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:07 crc kubenswrapper[4979]: I1204 12:03:07.997845 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.075009 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.093450 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.164664807 podStartE2EDuration="3.093432911s" podCreationTimestamp="2025-12-04 12:03:05 +0000 UTC" firstStartedPulling="2025-12-04 12:03:06.019700208 +0000 UTC m=+1210.293996012" lastFinishedPulling="2025-12-04 12:03:06.948468312 +0000 UTC m=+1211.222764116" observedRunningTime="2025-12-04 12:03:08.019089111 +0000 UTC m=+1212.293384925" watchObservedRunningTime="2025-12-04 12:03:08.093432911 +0000 UTC m=+1212.367728705" Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.098105 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.130690 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ac67-account-create-update-ztshn"] Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.133572 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.136552 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zxkmj"] Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.450503 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-375f-account-create-update-bh966"] Dec 04 12:03:08 crc kubenswrapper[4979]: I1204 12:03:08.631738 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-wl8cj"] Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.006759 4979 generic.go:334] "Generic (PLEG): container finished" podID="09068fde-06eb-4075-b4aa-78c8de781ced" containerID="01098447b1c91b182ef6963678cdbefbac832a1ebb6af0b9649ec51c2c66bb3b" exitCode=0 Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.006886 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ac67-account-create-update-ztshn" event={"ID":"09068fde-06eb-4075-b4aa-78c8de781ced","Type":"ContainerDied","Data":"01098447b1c91b182ef6963678cdbefbac832a1ebb6af0b9649ec51c2c66bb3b"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.007399 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ac67-account-create-update-ztshn" event={"ID":"09068fde-06eb-4075-b4aa-78c8de781ced","Type":"ContainerStarted","Data":"2be776e053547665f0a4e2ab8aa09b8088e19b546d076d4ffe52230c27f36e3d"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.008839 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-375f-account-create-update-bh966" event={"ID":"91e3fea1-7f8b-454c-9033-75cd30247761","Type":"ContainerStarted","Data":"b60e9bbefd212c20469dd92a8d6872954392efd8ce66ca080f0108e8f23c0598"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.008889 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-375f-account-create-update-bh966" event={"ID":"91e3fea1-7f8b-454c-9033-75cd30247761","Type":"ContainerStarted","Data":"3b0c1122f86659c72c1b565c3be57ae4dca74756472eba09bd9f73a27bb76104"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.010669 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wl8cj" event={"ID":"ad08b8b9-6e6c-40f3-9cf4-bb5165229869","Type":"ContainerStarted","Data":"905bf4fc0f5406fa0d2d6f9b94de151c0ca8b3a78f05d8bb25b374f9b58efd38"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.010702 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wl8cj" event={"ID":"ad08b8b9-6e6c-40f3-9cf4-bb5165229869","Type":"ContainerStarted","Data":"216d3ada309ce9eec42261ff7dbc94a055391617c0cd606b5358b1963a079007"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.013863 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zxkmj" event={"ID":"35e82e89-f20a-451f-b992-5b07d77b1e14","Type":"ContainerDied","Data":"a90821dd09be76982c32dd1c96a215f336b4994e444a92a4fa9851a9ea85a970"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.013822 4979 generic.go:334] "Generic (PLEG): container finished" podID="35e82e89-f20a-451f-b992-5b07d77b1e14" containerID="a90821dd09be76982c32dd1c96a215f336b4994e444a92a4fa9851a9ea85a970" exitCode=0 Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.013980 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zxkmj" event={"ID":"35e82e89-f20a-451f-b992-5b07d77b1e14","Type":"ContainerStarted","Data":"f5da922398f060c642abd20b66547e06cc9cc19b7b19969292d548b2176d9e9c"} Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.055351 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-wl8cj" podStartSLOduration=2.055332526 podStartE2EDuration="2.055332526s" podCreationTimestamp="2025-12-04 12:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:09.051559323 +0000 UTC m=+1213.325855127" watchObservedRunningTime="2025-12-04 12:03:09.055332526 +0000 UTC m=+1213.329628330" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.080663 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-375f-account-create-update-bh966" podStartSLOduration=2.080642394 podStartE2EDuration="2.080642394s" podCreationTimestamp="2025-12-04 12:03:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:09.071948788 +0000 UTC m=+1213.346244602" watchObservedRunningTime="2025-12-04 12:03:09.080642394 +0000 UTC m=+1213.354938198" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.182976 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-prxtm"] Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.213270 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-zqq2b"] Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.214844 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.229756 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zqq2b"] Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.256766 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.355721 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrj85\" (UniqueName: \"kubernetes.io/projected/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-kube-api-access-xrj85\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.355777 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-dns-svc\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.355804 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.355833 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-config\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.355895 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.456843 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrj85\" (UniqueName: \"kubernetes.io/projected/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-kube-api-access-xrj85\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.456888 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-dns-svc\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.456916 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.456944 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-config\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.456996 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.457837 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-dns-svc\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.457931 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.457985 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-config\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.459484 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.500654 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrj85\" (UniqueName: \"kubernetes.io/projected/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-kube-api-access-xrj85\") pod \"dnsmasq-dns-698758b865-zqq2b\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:09 crc kubenswrapper[4979]: I1204 12:03:09.553615 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.022129 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48","Type":"ContainerStarted","Data":"23f122939761be5f1b8c15a337a24423951f7eb2b8b6183e7f17f9e0cc30c1e6"} Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.024390 4979 generic.go:334] "Generic (PLEG): container finished" podID="91e3fea1-7f8b-454c-9033-75cd30247761" containerID="b60e9bbefd212c20469dd92a8d6872954392efd8ce66ca080f0108e8f23c0598" exitCode=0 Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.024470 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-375f-account-create-update-bh966" event={"ID":"91e3fea1-7f8b-454c-9033-75cd30247761","Type":"ContainerDied","Data":"b60e9bbefd212c20469dd92a8d6872954392efd8ce66ca080f0108e8f23c0598"} Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.026151 4979 generic.go:334] "Generic (PLEG): container finished" podID="ad08b8b9-6e6c-40f3-9cf4-bb5165229869" containerID="905bf4fc0f5406fa0d2d6f9b94de151c0ca8b3a78f05d8bb25b374f9b58efd38" exitCode=0 Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.026356 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerName="dnsmasq-dns" containerID="cri-o://6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041" gracePeriod=10 Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.026690 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wl8cj" event={"ID":"ad08b8b9-6e6c-40f3-9cf4-bb5165229869","Type":"ContainerDied","Data":"905bf4fc0f5406fa0d2d6f9b94de151c0ca8b3a78f05d8bb25b374f9b58efd38"} Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.060432 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zqq2b"] Dec 04 12:03:10 crc kubenswrapper[4979]: W1204 12:03:10.125653 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3c7923a_4a30_4260_ab5a_56d55fbab3e2.slice/crio-4445780ec18532b9d5a1cad9a33058367cdd4edd9468f5dc0513d4b561b08ebc WatchSource:0}: Error finding container 4445780ec18532b9d5a1cad9a33058367cdd4edd9468f5dc0513d4b561b08ebc: Status 404 returned error can't find the container with id 4445780ec18532b9d5a1cad9a33058367cdd4edd9468f5dc0513d4b561b08ebc Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.300877 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.301670 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.305980 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35e82e89-f20a-451f-b992-5b07d77b1e14" containerName="mariadb-database-create" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.306101 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="35e82e89-f20a-451f-b992-5b07d77b1e14" containerName="mariadb-database-create" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.306332 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="35e82e89-f20a-451f-b992-5b07d77b1e14" containerName="mariadb-database-create" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.311995 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.318280 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-kfk78" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.323678 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.324223 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.324607 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.337745 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474138 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e82e89-f20a-451f-b992-5b07d77b1e14-operator-scripts\") pod \"35e82e89-f20a-451f-b992-5b07d77b1e14\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474433 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ptk6\" (UniqueName: \"kubernetes.io/projected/35e82e89-f20a-451f-b992-5b07d77b1e14-kube-api-access-2ptk6\") pod \"35e82e89-f20a-451f-b992-5b07d77b1e14\" (UID: \"35e82e89-f20a-451f-b992-5b07d77b1e14\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474705 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-cache\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474820 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-lock\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474859 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474918 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.474941 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gqz7\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-kube-api-access-6gqz7\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.476182 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35e82e89-f20a-451f-b992-5b07d77b1e14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35e82e89-f20a-451f-b992-5b07d77b1e14" (UID: "35e82e89-f20a-451f-b992-5b07d77b1e14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.477623 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.479260 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35e82e89-f20a-451f-b992-5b07d77b1e14-kube-api-access-2ptk6" (OuterVolumeSpecName: "kube-api-access-2ptk6") pod "35e82e89-f20a-451f-b992-5b07d77b1e14" (UID: "35e82e89-f20a-451f-b992-5b07d77b1e14"). InnerVolumeSpecName "kube-api-access-2ptk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.486633 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.577875 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-config\") pod \"37734cd3-9b2b-47e0-85bb-3161d95229d3\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.577912 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-dns-svc\") pod \"37734cd3-9b2b-47e0-85bb-3161d95229d3\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.577972 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09068fde-06eb-4075-b4aa-78c8de781ced-operator-scripts\") pod \"09068fde-06eb-4075-b4aa-78c8de781ced\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.577998 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m7nh\" (UniqueName: \"kubernetes.io/projected/09068fde-06eb-4075-b4aa-78c8de781ced-kube-api-access-7m7nh\") pod \"09068fde-06eb-4075-b4aa-78c8de781ced\" (UID: \"09068fde-06eb-4075-b4aa-78c8de781ced\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578057 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-ovsdbserver-nb\") pod \"37734cd3-9b2b-47e0-85bb-3161d95229d3\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578104 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wt6c\" (UniqueName: \"kubernetes.io/projected/37734cd3-9b2b-47e0-85bb-3161d95229d3-kube-api-access-4wt6c\") pod \"37734cd3-9b2b-47e0-85bb-3161d95229d3\" (UID: \"37734cd3-9b2b-47e0-85bb-3161d95229d3\") " Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578289 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-lock\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578336 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578393 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578416 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gqz7\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-kube-api-access-6gqz7\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578485 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-cache\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578582 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ptk6\" (UniqueName: \"kubernetes.io/projected/35e82e89-f20a-451f-b992-5b07d77b1e14-kube-api-access-2ptk6\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.578596 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35e82e89-f20a-451f-b992-5b07d77b1e14-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.579027 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-cache\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.579264 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.579322 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-lock\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.579350 4979 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.579404 4979 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.579452 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift podName:4eafdbc5-ec8d-4540-a82f-b628ad83fb08 nodeName:}" failed. No retries permitted until 2025-12-04 12:03:11.079435955 +0000 UTC m=+1215.353731759 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift") pod "swift-storage-0" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08") : configmap "swift-ring-files" not found Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.579497 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09068fde-06eb-4075-b4aa-78c8de781ced-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09068fde-06eb-4075-b4aa-78c8de781ced" (UID: "09068fde-06eb-4075-b4aa-78c8de781ced"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.588626 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37734cd3-9b2b-47e0-85bb-3161d95229d3-kube-api-access-4wt6c" (OuterVolumeSpecName: "kube-api-access-4wt6c") pod "37734cd3-9b2b-47e0-85bb-3161d95229d3" (UID: "37734cd3-9b2b-47e0-85bb-3161d95229d3"). InnerVolumeSpecName "kube-api-access-4wt6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.590329 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09068fde-06eb-4075-b4aa-78c8de781ced-kube-api-access-7m7nh" (OuterVolumeSpecName: "kube-api-access-7m7nh") pod "09068fde-06eb-4075-b4aa-78c8de781ced" (UID: "09068fde-06eb-4075-b4aa-78c8de781ced"). InnerVolumeSpecName "kube-api-access-7m7nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.600201 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gqz7\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-kube-api-access-6gqz7\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.617143 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.623672 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-config" (OuterVolumeSpecName: "config") pod "37734cd3-9b2b-47e0-85bb-3161d95229d3" (UID: "37734cd3-9b2b-47e0-85bb-3161d95229d3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.624996 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37734cd3-9b2b-47e0-85bb-3161d95229d3" (UID: "37734cd3-9b2b-47e0-85bb-3161d95229d3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.627594 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37734cd3-9b2b-47e0-85bb-3161d95229d3" (UID: "37734cd3-9b2b-47e0-85bb-3161d95229d3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.679851 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09068fde-06eb-4075-b4aa-78c8de781ced-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.679883 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m7nh\" (UniqueName: \"kubernetes.io/projected/09068fde-06eb-4075-b4aa-78c8de781ced-kube-api-access-7m7nh\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.679894 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.679902 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wt6c\" (UniqueName: \"kubernetes.io/projected/37734cd3-9b2b-47e0-85bb-3161d95229d3-kube-api-access-4wt6c\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.679910 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.679918 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37734cd3-9b2b-47e0-85bb-3161d95229d3-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.952968 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-t2bg8"] Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.953636 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerName="init" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.953663 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerName="init" Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.953694 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerName="dnsmasq-dns" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.953703 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerName="dnsmasq-dns" Dec 04 12:03:10 crc kubenswrapper[4979]: E1204 12:03:10.953717 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09068fde-06eb-4075-b4aa-78c8de781ced" containerName="mariadb-account-create-update" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.953725 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="09068fde-06eb-4075-b4aa-78c8de781ced" containerName="mariadb-account-create-update" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.953939 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="09068fde-06eb-4075-b4aa-78c8de781ced" containerName="mariadb-account-create-update" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.953966 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerName="dnsmasq-dns" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.954653 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.955962 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.956272 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.956524 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 04 12:03:10 crc kubenswrapper[4979]: I1204 12:03:10.963174 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t2bg8"] Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.034406 4979 generic.go:334] "Generic (PLEG): container finished" podID="37734cd3-9b2b-47e0-85bb-3161d95229d3" containerID="6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041" exitCode=0 Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.034467 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" event={"ID":"37734cd3-9b2b-47e0-85bb-3161d95229d3","Type":"ContainerDied","Data":"6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041"} Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.034492 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" event={"ID":"37734cd3-9b2b-47e0-85bb-3161d95229d3","Type":"ContainerDied","Data":"8e721dc08e1b1e75e17d9779c958399c37da861bcb668f081e7da2ba01c290b6"} Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.034508 4979 scope.go:117] "RemoveContainer" containerID="6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.034634 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7fd796d7df-prxtm" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.037455 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zxkmj" event={"ID":"35e82e89-f20a-451f-b992-5b07d77b1e14","Type":"ContainerDied","Data":"f5da922398f060c642abd20b66547e06cc9cc19b7b19969292d548b2176d9e9c"} Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.037489 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5da922398f060c642abd20b66547e06cc9cc19b7b19969292d548b2176d9e9c" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.037512 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zxkmj" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.037917 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-scripts\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.037966 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-combined-ca-bundle\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038031 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-ring-data-devices\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038069 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-swiftconf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038145 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v2vf\" (UniqueName: \"kubernetes.io/projected/e14c1da3-5204-4d84-94fd-039b1cef8af3-kube-api-access-4v2vf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038250 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e14c1da3-5204-4d84-94fd-039b1cef8af3-etc-swift\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038271 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-dispersionconf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038916 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ac67-account-create-update-ztshn" event={"ID":"09068fde-06eb-4075-b4aa-78c8de781ced","Type":"ContainerDied","Data":"2be776e053547665f0a4e2ab8aa09b8088e19b546d076d4ffe52230c27f36e3d"} Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038966 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ac67-account-create-update-ztshn" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.038972 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2be776e053547665f0a4e2ab8aa09b8088e19b546d076d4ffe52230c27f36e3d" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.040088 4979 generic.go:334] "Generic (PLEG): container finished" podID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerID="8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5" exitCode=0 Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.040952 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zqq2b" event={"ID":"b3c7923a-4a30-4260-ab5a-56d55fbab3e2","Type":"ContainerDied","Data":"8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5"} Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.040979 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zqq2b" event={"ID":"b3c7923a-4a30-4260-ab5a-56d55fbab3e2","Type":"ContainerStarted","Data":"4445780ec18532b9d5a1cad9a33058367cdd4edd9468f5dc0513d4b561b08ebc"} Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.092234 4979 scope.go:117] "RemoveContainer" containerID="8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.139247 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e14c1da3-5204-4d84-94fd-039b1cef8af3-etc-swift\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.139286 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-dispersionconf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140499 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-scripts\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140550 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-combined-ca-bundle\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140640 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-ring-data-devices\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140683 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-swiftconf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140720 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e14c1da3-5204-4d84-94fd-039b1cef8af3-etc-swift\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140730 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.140793 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v2vf\" (UniqueName: \"kubernetes.io/projected/e14c1da3-5204-4d84-94fd-039b1cef8af3-kube-api-access-4v2vf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: E1204 12:03:11.140846 4979 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 12:03:11 crc kubenswrapper[4979]: E1204 12:03:11.140860 4979 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 12:03:11 crc kubenswrapper[4979]: E1204 12:03:11.140906 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift podName:4eafdbc5-ec8d-4540-a82f-b628ad83fb08 nodeName:}" failed. No retries permitted until 2025-12-04 12:03:12.14088884 +0000 UTC m=+1216.415184644 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift") pod "swift-storage-0" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08") : configmap "swift-ring-files" not found Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.141738 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-scripts\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.143411 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-ring-data-devices\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.148919 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-dispersionconf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.149359 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-combined-ca-bundle\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.164954 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v2vf\" (UniqueName: \"kubernetes.io/projected/e14c1da3-5204-4d84-94fd-039b1cef8af3-kube-api-access-4v2vf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.165557 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-swiftconf\") pod \"swift-ring-rebalance-t2bg8\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.186334 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-prxtm"] Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.187440 4979 scope.go:117] "RemoveContainer" containerID="6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041" Dec 04 12:03:11 crc kubenswrapper[4979]: E1204 12:03:11.187862 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041\": container with ID starting with 6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041 not found: ID does not exist" containerID="6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.187918 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041"} err="failed to get container status \"6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041\": rpc error: code = NotFound desc = could not find container \"6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041\": container with ID starting with 6f8dee4acafa76e345d3e570fdcd47a6a4e889ca598f4d88b39dc0e29aa8b041 not found: ID does not exist" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.187945 4979 scope.go:117] "RemoveContainer" containerID="8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51" Dec 04 12:03:11 crc kubenswrapper[4979]: E1204 12:03:11.188170 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51\": container with ID starting with 8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51 not found: ID does not exist" containerID="8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.188191 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51"} err="failed to get container status \"8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51\": rpc error: code = NotFound desc = could not find container \"8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51\": container with ID starting with 8ae64e643fa9394d4b48381866c578939b4b4ec34efabd7df0432098a99a0b51 not found: ID does not exist" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.192319 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7fd796d7df-prxtm"] Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.275741 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.510117 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.625021 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.651124 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n2xt\" (UniqueName: \"kubernetes.io/projected/91e3fea1-7f8b-454c-9033-75cd30247761-kube-api-access-4n2xt\") pod \"91e3fea1-7f8b-454c-9033-75cd30247761\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.651194 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91e3fea1-7f8b-454c-9033-75cd30247761-operator-scripts\") pod \"91e3fea1-7f8b-454c-9033-75cd30247761\" (UID: \"91e3fea1-7f8b-454c-9033-75cd30247761\") " Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.652147 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/91e3fea1-7f8b-454c-9033-75cd30247761-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "91e3fea1-7f8b-454c-9033-75cd30247761" (UID: "91e3fea1-7f8b-454c-9033-75cd30247761"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.668756 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e3fea1-7f8b-454c-9033-75cd30247761-kube-api-access-4n2xt" (OuterVolumeSpecName: "kube-api-access-4n2xt") pod "91e3fea1-7f8b-454c-9033-75cd30247761" (UID: "91e3fea1-7f8b-454c-9033-75cd30247761"). InnerVolumeSpecName "kube-api-access-4n2xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.752640 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-operator-scripts\") pod \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.752788 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrn28\" (UniqueName: \"kubernetes.io/projected/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-kube-api-access-wrn28\") pod \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\" (UID: \"ad08b8b9-6e6c-40f3-9cf4-bb5165229869\") " Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.753125 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad08b8b9-6e6c-40f3-9cf4-bb5165229869" (UID: "ad08b8b9-6e6c-40f3-9cf4-bb5165229869"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.753512 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.753534 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n2xt\" (UniqueName: \"kubernetes.io/projected/91e3fea1-7f8b-454c-9033-75cd30247761-kube-api-access-4n2xt\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.753548 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/91e3fea1-7f8b-454c-9033-75cd30247761-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.757146 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-kube-api-access-wrn28" (OuterVolumeSpecName: "kube-api-access-wrn28") pod "ad08b8b9-6e6c-40f3-9cf4-bb5165229869" (UID: "ad08b8b9-6e6c-40f3-9cf4-bb5165229869"). InnerVolumeSpecName "kube-api-access-wrn28". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.854781 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrn28\" (UniqueName: \"kubernetes.io/projected/ad08b8b9-6e6c-40f3-9cf4-bb5165229869-kube-api-access-wrn28\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:11 crc kubenswrapper[4979]: I1204 12:03:11.919507 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-t2bg8"] Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.047487 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-375f-account-create-update-bh966" event={"ID":"91e3fea1-7f8b-454c-9033-75cd30247761","Type":"ContainerDied","Data":"3b0c1122f86659c72c1b565c3be57ae4dca74756472eba09bd9f73a27bb76104"} Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.047520 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b0c1122f86659c72c1b565c3be57ae4dca74756472eba09bd9f73a27bb76104" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.047897 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-375f-account-create-update-bh966" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.049563 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-wl8cj" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.049578 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-wl8cj" event={"ID":"ad08b8b9-6e6c-40f3-9cf4-bb5165229869","Type":"ContainerDied","Data":"216d3ada309ce9eec42261ff7dbc94a055391617c0cd606b5358b1963a079007"} Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.049631 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="216d3ada309ce9eec42261ff7dbc94a055391617c0cd606b5358b1963a079007" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.051219 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t2bg8" event={"ID":"e14c1da3-5204-4d84-94fd-039b1cef8af3","Type":"ContainerStarted","Data":"e002b60f895518acc573949449b33f96e4d7b791f83fbbe60399e9dd7784afa0"} Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.160356 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:12 crc kubenswrapper[4979]: E1204 12:03:12.160488 4979 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 12:03:12 crc kubenswrapper[4979]: E1204 12:03:12.160509 4979 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 12:03:12 crc kubenswrapper[4979]: E1204 12:03:12.160555 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift podName:4eafdbc5-ec8d-4540-a82f-b628ad83fb08 nodeName:}" failed. No retries permitted until 2025-12-04 12:03:14.160540265 +0000 UTC m=+1218.434836069 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift") pod "swift-storage-0" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08") : configmap "swift-ring-files" not found Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.207404 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37734cd3-9b2b-47e0-85bb-3161d95229d3" path="/var/lib/kubelet/pods/37734cd3-9b2b-47e0-85bb-3161d95229d3/volumes" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.845980 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-b2dt6"] Dec 04 12:03:12 crc kubenswrapper[4979]: E1204 12:03:12.846698 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad08b8b9-6e6c-40f3-9cf4-bb5165229869" containerName="mariadb-database-create" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.846715 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad08b8b9-6e6c-40f3-9cf4-bb5165229869" containerName="mariadb-database-create" Dec 04 12:03:12 crc kubenswrapper[4979]: E1204 12:03:12.846731 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e3fea1-7f8b-454c-9033-75cd30247761" containerName="mariadb-account-create-update" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.846737 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e3fea1-7f8b-454c-9033-75cd30247761" containerName="mariadb-account-create-update" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.846919 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e3fea1-7f8b-454c-9033-75cd30247761" containerName="mariadb-account-create-update" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.846941 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad08b8b9-6e6c-40f3-9cf4-bb5165229869" containerName="mariadb-database-create" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.847534 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.868858 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-b2dt6"] Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.872364 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmdxq\" (UniqueName: \"kubernetes.io/projected/f250de55-be06-43c1-9366-b3575dc45061-kube-api-access-tmdxq\") pod \"glance-db-create-b2dt6\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.872404 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f250de55-be06-43c1-9366-b3575dc45061-operator-scripts\") pod \"glance-db-create-b2dt6\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.945094 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-2ebe-account-create-update-tngwj"] Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.946069 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.948228 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.962607 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2ebe-account-create-update-tngwj"] Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.979572 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmdxq\" (UniqueName: \"kubernetes.io/projected/f250de55-be06-43c1-9366-b3575dc45061-kube-api-access-tmdxq\") pod \"glance-db-create-b2dt6\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.979636 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f250de55-be06-43c1-9366-b3575dc45061-operator-scripts\") pod \"glance-db-create-b2dt6\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:12 crc kubenswrapper[4979]: I1204 12:03:12.980378 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f250de55-be06-43c1-9366-b3575dc45061-operator-scripts\") pod \"glance-db-create-b2dt6\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.003561 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmdxq\" (UniqueName: \"kubernetes.io/projected/f250de55-be06-43c1-9366-b3575dc45061-kube-api-access-tmdxq\") pod \"glance-db-create-b2dt6\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.080978 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7ptk\" (UniqueName: \"kubernetes.io/projected/29b7139c-f1fc-4637-8288-053f85ab6202-kube-api-access-t7ptk\") pod \"glance-2ebe-account-create-update-tngwj\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.081075 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29b7139c-f1fc-4637-8288-053f85ab6202-operator-scripts\") pod \"glance-2ebe-account-create-update-tngwj\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.168227 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.183091 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7ptk\" (UniqueName: \"kubernetes.io/projected/29b7139c-f1fc-4637-8288-053f85ab6202-kube-api-access-t7ptk\") pod \"glance-2ebe-account-create-update-tngwj\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.183217 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29b7139c-f1fc-4637-8288-053f85ab6202-operator-scripts\") pod \"glance-2ebe-account-create-update-tngwj\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.184148 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29b7139c-f1fc-4637-8288-053f85ab6202-operator-scripts\") pod \"glance-2ebe-account-create-update-tngwj\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.206120 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7ptk\" (UniqueName: \"kubernetes.io/projected/29b7139c-f1fc-4637-8288-053f85ab6202-kube-api-access-t7ptk\") pod \"glance-2ebe-account-create-update-tngwj\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.260916 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.627189 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-b2dt6"] Dec 04 12:03:13 crc kubenswrapper[4979]: W1204 12:03:13.636764 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250de55_be06_43c1_9366_b3575dc45061.slice/crio-34d341630db0ed920cb0a48ac82967f4e7518d4dc300f78cc69c36e88d944c30 WatchSource:0}: Error finding container 34d341630db0ed920cb0a48ac82967f4e7518d4dc300f78cc69c36e88d944c30: Status 404 returned error can't find the container with id 34d341630db0ed920cb0a48ac82967f4e7518d4dc300f78cc69c36e88d944c30 Dec 04 12:03:13 crc kubenswrapper[4979]: I1204 12:03:13.731070 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-2ebe-account-create-update-tngwj"] Dec 04 12:03:13 crc kubenswrapper[4979]: W1204 12:03:13.735058 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29b7139c_f1fc_4637_8288_053f85ab6202.slice/crio-384a9512022b45206ef0885030c79e129c325d5c33b70158c06c09e99bd05587 WatchSource:0}: Error finding container 384a9512022b45206ef0885030c79e129c325d5c33b70158c06c09e99bd05587: Status 404 returned error can't find the container with id 384a9512022b45206ef0885030c79e129c325d5c33b70158c06c09e99bd05587 Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.072248 4979 generic.go:334] "Generic (PLEG): container finished" podID="f250de55-be06-43c1-9366-b3575dc45061" containerID="14b0d7a646aefbe6dd898dc529926522899476170a2a82286a54b26b858ad281" exitCode=0 Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.072616 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-b2dt6" event={"ID":"f250de55-be06-43c1-9366-b3575dc45061","Type":"ContainerDied","Data":"14b0d7a646aefbe6dd898dc529926522899476170a2a82286a54b26b858ad281"} Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.072647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-b2dt6" event={"ID":"f250de55-be06-43c1-9366-b3575dc45061","Type":"ContainerStarted","Data":"34d341630db0ed920cb0a48ac82967f4e7518d4dc300f78cc69c36e88d944c30"} Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.075235 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zqq2b" event={"ID":"b3c7923a-4a30-4260-ab5a-56d55fbab3e2","Type":"ContainerStarted","Data":"3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b"} Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.075432 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.077131 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2ebe-account-create-update-tngwj" event={"ID":"29b7139c-f1fc-4637-8288-053f85ab6202","Type":"ContainerStarted","Data":"20abc3d60657fd8e86970f06db37aaca534a66acbeede0e238a1dc8d387e98fd"} Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.077154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2ebe-account-create-update-tngwj" event={"ID":"29b7139c-f1fc-4637-8288-053f85ab6202","Type":"ContainerStarted","Data":"384a9512022b45206ef0885030c79e129c325d5c33b70158c06c09e99bd05587"} Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.106022 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-zqq2b" podStartSLOduration=5.105964362 podStartE2EDuration="5.105964362s" podCreationTimestamp="2025-12-04 12:03:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:14.104518802 +0000 UTC m=+1218.378814606" watchObservedRunningTime="2025-12-04 12:03:14.105964362 +0000 UTC m=+1218.380260166" Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.128213 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-2ebe-account-create-update-tngwj" podStartSLOduration=2.1281891059999998 podStartE2EDuration="2.128189106s" podCreationTimestamp="2025-12-04 12:03:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:14.117911357 +0000 UTC m=+1218.392207161" watchObservedRunningTime="2025-12-04 12:03:14.128189106 +0000 UTC m=+1218.402484920" Dec 04 12:03:14 crc kubenswrapper[4979]: I1204 12:03:14.210231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:14 crc kubenswrapper[4979]: E1204 12:03:14.210453 4979 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 12:03:14 crc kubenswrapper[4979]: E1204 12:03:14.210473 4979 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 12:03:14 crc kubenswrapper[4979]: E1204 12:03:14.210526 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift podName:4eafdbc5-ec8d-4540-a82f-b628ad83fb08 nodeName:}" failed. No retries permitted until 2025-12-04 12:03:18.210505982 +0000 UTC m=+1222.484801786 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift") pod "swift-storage-0" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08") : configmap "swift-ring-files" not found Dec 04 12:03:15 crc kubenswrapper[4979]: I1204 12:03:15.084961 4979 generic.go:334] "Generic (PLEG): container finished" podID="29b7139c-f1fc-4637-8288-053f85ab6202" containerID="20abc3d60657fd8e86970f06db37aaca534a66acbeede0e238a1dc8d387e98fd" exitCode=0 Dec 04 12:03:15 crc kubenswrapper[4979]: I1204 12:03:15.085063 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2ebe-account-create-update-tngwj" event={"ID":"29b7139c-f1fc-4637-8288-053f85ab6202","Type":"ContainerDied","Data":"20abc3d60657fd8e86970f06db37aaca534a66acbeede0e238a1dc8d387e98fd"} Dec 04 12:03:15 crc kubenswrapper[4979]: I1204 12:03:15.124606 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:16 crc kubenswrapper[4979]: I1204 12:03:16.936139 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:16 crc kubenswrapper[4979]: I1204 12:03:16.960704 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:16 crc kubenswrapper[4979]: I1204 12:03:16.999166 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7ptk\" (UniqueName: \"kubernetes.io/projected/29b7139c-f1fc-4637-8288-053f85ab6202-kube-api-access-t7ptk\") pod \"29b7139c-f1fc-4637-8288-053f85ab6202\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " Dec 04 12:03:16 crc kubenswrapper[4979]: I1204 12:03:16.999346 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f250de55-be06-43c1-9366-b3575dc45061-operator-scripts\") pod \"f250de55-be06-43c1-9366-b3575dc45061\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " Dec 04 12:03:16 crc kubenswrapper[4979]: I1204 12:03:16.999453 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29b7139c-f1fc-4637-8288-053f85ab6202-operator-scripts\") pod \"29b7139c-f1fc-4637-8288-053f85ab6202\" (UID: \"29b7139c-f1fc-4637-8288-053f85ab6202\") " Dec 04 12:03:16 crc kubenswrapper[4979]: I1204 12:03:16.999619 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmdxq\" (UniqueName: \"kubernetes.io/projected/f250de55-be06-43c1-9366-b3575dc45061-kube-api-access-tmdxq\") pod \"f250de55-be06-43c1-9366-b3575dc45061\" (UID: \"f250de55-be06-43c1-9366-b3575dc45061\") " Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:16.999997 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29b7139c-f1fc-4637-8288-053f85ab6202-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "29b7139c-f1fc-4637-8288-053f85ab6202" (UID: "29b7139c-f1fc-4637-8288-053f85ab6202"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.000126 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f250de55-be06-43c1-9366-b3575dc45061-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f250de55-be06-43c1-9366-b3575dc45061" (UID: "f250de55-be06-43c1-9366-b3575dc45061"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.000187 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/29b7139c-f1fc-4637-8288-053f85ab6202-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.004345 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f250de55-be06-43c1-9366-b3575dc45061-kube-api-access-tmdxq" (OuterVolumeSpecName: "kube-api-access-tmdxq") pod "f250de55-be06-43c1-9366-b3575dc45061" (UID: "f250de55-be06-43c1-9366-b3575dc45061"). InnerVolumeSpecName "kube-api-access-tmdxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.005278 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29b7139c-f1fc-4637-8288-053f85ab6202-kube-api-access-t7ptk" (OuterVolumeSpecName: "kube-api-access-t7ptk") pod "29b7139c-f1fc-4637-8288-053f85ab6202" (UID: "29b7139c-f1fc-4637-8288-053f85ab6202"). InnerVolumeSpecName "kube-api-access-t7ptk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.101337 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmdxq\" (UniqueName: \"kubernetes.io/projected/f250de55-be06-43c1-9366-b3575dc45061-kube-api-access-tmdxq\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.101382 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7ptk\" (UniqueName: \"kubernetes.io/projected/29b7139c-f1fc-4637-8288-053f85ab6202-kube-api-access-t7ptk\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.101396 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f250de55-be06-43c1-9366-b3575dc45061-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.115067 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-2ebe-account-create-update-tngwj" event={"ID":"29b7139c-f1fc-4637-8288-053f85ab6202","Type":"ContainerDied","Data":"384a9512022b45206ef0885030c79e129c325d5c33b70158c06c09e99bd05587"} Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.115122 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="384a9512022b45206ef0885030c79e129c325d5c33b70158c06c09e99bd05587" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.115197 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-2ebe-account-create-update-tngwj" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.117905 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t2bg8" event={"ID":"e14c1da3-5204-4d84-94fd-039b1cef8af3","Type":"ContainerStarted","Data":"3f0243e98bb73e51c6d7086f697067a94b8297badef0bd16f3868920634c8218"} Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.119567 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-b2dt6" event={"ID":"f250de55-be06-43c1-9366-b3575dc45061","Type":"ContainerDied","Data":"34d341630db0ed920cb0a48ac82967f4e7518d4dc300f78cc69c36e88d944c30"} Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.119611 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34d341630db0ed920cb0a48ac82967f4e7518d4dc300f78cc69c36e88d944c30" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.119646 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-b2dt6" Dec 04 12:03:17 crc kubenswrapper[4979]: I1204 12:03:17.137477 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-t2bg8" podStartSLOduration=2.260625423 podStartE2EDuration="7.137456837s" podCreationTimestamp="2025-12-04 12:03:10 +0000 UTC" firstStartedPulling="2025-12-04 12:03:11.917157282 +0000 UTC m=+1216.191453086" lastFinishedPulling="2025-12-04 12:03:16.793988696 +0000 UTC m=+1221.068284500" observedRunningTime="2025-12-04 12:03:17.133719736 +0000 UTC m=+1221.408015540" watchObservedRunningTime="2025-12-04 12:03:17.137456837 +0000 UTC m=+1221.411752661" Dec 04 12:03:18 crc kubenswrapper[4979]: I1204 12:03:18.219631 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:18 crc kubenswrapper[4979]: E1204 12:03:18.219804 4979 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 12:03:18 crc kubenswrapper[4979]: E1204 12:03:18.219977 4979 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 12:03:18 crc kubenswrapper[4979]: E1204 12:03:18.220026 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift podName:4eafdbc5-ec8d-4540-a82f-b628ad83fb08 nodeName:}" failed. No retries permitted until 2025-12-04 12:03:26.22000947 +0000 UTC m=+1230.494305274 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift") pod "swift-storage-0" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08") : configmap "swift-ring-files" not found Dec 04 12:03:19 crc kubenswrapper[4979]: I1204 12:03:19.555553 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:03:19 crc kubenswrapper[4979]: I1204 12:03:19.617644 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kpn2l"] Dec 04 12:03:19 crc kubenswrapper[4979]: I1204 12:03:19.618067 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="dnsmasq-dns" containerID="cri-o://f1c694f83174118748120db5cc618d6289efeb90b3ddca198eb54de151f03ee5" gracePeriod=10 Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.123804 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.146866 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6666948-dde4-490a-93af-d5494b1e8243" containerID="f1c694f83174118748120db5cc618d6289efeb90b3ddca198eb54de151f03ee5" exitCode=0 Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.146945 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" event={"ID":"a6666948-dde4-490a-93af-d5494b1e8243","Type":"ContainerDied","Data":"f1c694f83174118748120db5cc618d6289efeb90b3ddca198eb54de151f03ee5"} Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.560382 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.590777 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.665109 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-dns-svc\") pod \"a6666948-dde4-490a-93af-d5494b1e8243\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.665152 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-sb\") pod \"a6666948-dde4-490a-93af-d5494b1e8243\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.665186 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-config\") pod \"a6666948-dde4-490a-93af-d5494b1e8243\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.665276 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5qvz\" (UniqueName: \"kubernetes.io/projected/a6666948-dde4-490a-93af-d5494b1e8243-kube-api-access-p5qvz\") pod \"a6666948-dde4-490a-93af-d5494b1e8243\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.665412 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-nb\") pod \"a6666948-dde4-490a-93af-d5494b1e8243\" (UID: \"a6666948-dde4-490a-93af-d5494b1e8243\") " Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.673046 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6666948-dde4-490a-93af-d5494b1e8243-kube-api-access-p5qvz" (OuterVolumeSpecName: "kube-api-access-p5qvz") pod "a6666948-dde4-490a-93af-d5494b1e8243" (UID: "a6666948-dde4-490a-93af-d5494b1e8243"). InnerVolumeSpecName "kube-api-access-p5qvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.712505 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a6666948-dde4-490a-93af-d5494b1e8243" (UID: "a6666948-dde4-490a-93af-d5494b1e8243"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.721672 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-config" (OuterVolumeSpecName: "config") pod "a6666948-dde4-490a-93af-d5494b1e8243" (UID: "a6666948-dde4-490a-93af-d5494b1e8243"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.725876 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6666948-dde4-490a-93af-d5494b1e8243" (UID: "a6666948-dde4-490a-93af-d5494b1e8243"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.727533 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a6666948-dde4-490a-93af-d5494b1e8243" (UID: "a6666948-dde4-490a-93af-d5494b1e8243"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.768944 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.768978 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.768986 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.768995 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6666948-dde4-490a-93af-d5494b1e8243-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:20 crc kubenswrapper[4979]: I1204 12:03:20.769005 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5qvz\" (UniqueName: \"kubernetes.io/projected/a6666948-dde4-490a-93af-d5494b1e8243-kube-api-access-p5qvz\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:21 crc kubenswrapper[4979]: I1204 12:03:21.160324 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" event={"ID":"a6666948-dde4-490a-93af-d5494b1e8243","Type":"ContainerDied","Data":"5d22631616b602de8536d56ef86994da40474e7622c6049f34e2d2d4a8a21fc1"} Dec 04 12:03:21 crc kubenswrapper[4979]: I1204 12:03:21.160730 4979 scope.go:117] "RemoveContainer" containerID="f1c694f83174118748120db5cc618d6289efeb90b3ddca198eb54de151f03ee5" Dec 04 12:03:21 crc kubenswrapper[4979]: I1204 12:03:21.160396 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-kpn2l" Dec 04 12:03:21 crc kubenswrapper[4979]: I1204 12:03:21.182495 4979 scope.go:117] "RemoveContainer" containerID="cc35a467283186803a3147f7ae91b7b31c57a2058c85bd55b0f75ef4ba36ebf2" Dec 04 12:03:21 crc kubenswrapper[4979]: I1204 12:03:21.221147 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kpn2l"] Dec 04 12:03:21 crc kubenswrapper[4979]: I1204 12:03:21.233081 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-kpn2l"] Dec 04 12:03:22 crc kubenswrapper[4979]: I1204 12:03:22.213947 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6666948-dde4-490a-93af-d5494b1e8243" path="/var/lib/kubelet/pods/a6666948-dde4-490a-93af-d5494b1e8243/volumes" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.096859 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-8kdt5"] Dec 04 12:03:23 crc kubenswrapper[4979]: E1204 12:03:23.097716 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="dnsmasq-dns" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.097741 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="dnsmasq-dns" Dec 04 12:03:23 crc kubenswrapper[4979]: E1204 12:03:23.097762 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29b7139c-f1fc-4637-8288-053f85ab6202" containerName="mariadb-account-create-update" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.097771 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="29b7139c-f1fc-4637-8288-053f85ab6202" containerName="mariadb-account-create-update" Dec 04 12:03:23 crc kubenswrapper[4979]: E1204 12:03:23.097790 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f250de55-be06-43c1-9366-b3575dc45061" containerName="mariadb-database-create" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.097797 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f250de55-be06-43c1-9366-b3575dc45061" containerName="mariadb-database-create" Dec 04 12:03:23 crc kubenswrapper[4979]: E1204 12:03:23.097821 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="init" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.097828 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="init" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.098063 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6666948-dde4-490a-93af-d5494b1e8243" containerName="dnsmasq-dns" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.098077 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f250de55-be06-43c1-9366-b3575dc45061" containerName="mariadb-database-create" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.098087 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="29b7139c-f1fc-4637-8288-053f85ab6202" containerName="mariadb-account-create-update" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.098875 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.100657 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ksp64" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.101142 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.114992 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8kdt5"] Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.212868 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-combined-ca-bundle\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.212944 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-db-sync-config-data\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.213251 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4849\" (UniqueName: \"kubernetes.io/projected/f4f85b75-ee47-4c71-822c-e30cd4f3492c-kube-api-access-h4849\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.213339 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-config-data\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.315002 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4849\" (UniqueName: \"kubernetes.io/projected/f4f85b75-ee47-4c71-822c-e30cd4f3492c-kube-api-access-h4849\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.315054 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-config-data\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.315146 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-combined-ca-bundle\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.315174 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-db-sync-config-data\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.323030 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-db-sync-config-data\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.323081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-combined-ca-bundle\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.323500 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-config-data\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.331661 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4849\" (UniqueName: \"kubernetes.io/projected/f4f85b75-ee47-4c71-822c-e30cd4f3492c-kube-api-access-h4849\") pod \"glance-db-sync-8kdt5\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:23 crc kubenswrapper[4979]: I1204 12:03:23.418254 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8kdt5" Dec 04 12:03:24 crc kubenswrapper[4979]: I1204 12:03:24.036822 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-8kdt5"] Dec 04 12:03:24 crc kubenswrapper[4979]: I1204 12:03:24.188466 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8kdt5" event={"ID":"f4f85b75-ee47-4c71-822c-e30cd4f3492c","Type":"ContainerStarted","Data":"03d55b25b66ba9dd2d4cfb5fab327d07adc05fda38777b816d5f4be3d200fe8d"} Dec 04 12:03:26 crc kubenswrapper[4979]: I1204 12:03:26.267103 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:26 crc kubenswrapper[4979]: E1204 12:03:26.267353 4979 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 04 12:03:26 crc kubenswrapper[4979]: E1204 12:03:26.269166 4979 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 04 12:03:26 crc kubenswrapper[4979]: E1204 12:03:26.269427 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift podName:4eafdbc5-ec8d-4540-a82f-b628ad83fb08 nodeName:}" failed. No retries permitted until 2025-12-04 12:03:42.269400503 +0000 UTC m=+1246.543696337 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift") pod "swift-storage-0" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08") : configmap "swift-ring-files" not found Dec 04 12:03:27 crc kubenswrapper[4979]: I1204 12:03:27.210618 4979 generic.go:334] "Generic (PLEG): container finished" podID="e14c1da3-5204-4d84-94fd-039b1cef8af3" containerID="3f0243e98bb73e51c6d7086f697067a94b8297badef0bd16f3868920634c8218" exitCode=0 Dec 04 12:03:27 crc kubenswrapper[4979]: I1204 12:03:27.210676 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t2bg8" event={"ID":"e14c1da3-5204-4d84-94fd-039b1cef8af3","Type":"ContainerDied","Data":"3f0243e98bb73e51c6d7086f697067a94b8297badef0bd16f3868920634c8218"} Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.603857 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.711978 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-combined-ca-bundle\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.712053 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e14c1da3-5204-4d84-94fd-039b1cef8af3-etc-swift\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.712149 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-dispersionconf\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.712216 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-ring-data-devices\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.712381 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-scripts\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.712430 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4v2vf\" (UniqueName: \"kubernetes.io/projected/e14c1da3-5204-4d84-94fd-039b1cef8af3-kube-api-access-4v2vf\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.712482 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-swiftconf\") pod \"e14c1da3-5204-4d84-94fd-039b1cef8af3\" (UID: \"e14c1da3-5204-4d84-94fd-039b1cef8af3\") " Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.713347 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.713763 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e14c1da3-5204-4d84-94fd-039b1cef8af3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.719440 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e14c1da3-5204-4d84-94fd-039b1cef8af3-kube-api-access-4v2vf" (OuterVolumeSpecName: "kube-api-access-4v2vf") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "kube-api-access-4v2vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.723342 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.739186 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.742540 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-scripts" (OuterVolumeSpecName: "scripts") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.744267 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e14c1da3-5204-4d84-94fd-039b1cef8af3" (UID: "e14c1da3-5204-4d84-94fd-039b1cef8af3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814226 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814252 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4v2vf\" (UniqueName: \"kubernetes.io/projected/e14c1da3-5204-4d84-94fd-039b1cef8af3-kube-api-access-4v2vf\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814289 4979 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814312 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814321 4979 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e14c1da3-5204-4d84-94fd-039b1cef8af3-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814328 4979 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e14c1da3-5204-4d84-94fd-039b1cef8af3-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:28 crc kubenswrapper[4979]: I1204 12:03:28.814337 4979 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e14c1da3-5204-4d84-94fd-039b1cef8af3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:29 crc kubenswrapper[4979]: I1204 12:03:29.233574 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-t2bg8" Dec 04 12:03:29 crc kubenswrapper[4979]: I1204 12:03:29.234455 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-t2bg8" event={"ID":"e14c1da3-5204-4d84-94fd-039b1cef8af3","Type":"ContainerDied","Data":"e002b60f895518acc573949449b33f96e4d7b791f83fbbe60399e9dd7784afa0"} Dec 04 12:03:29 crc kubenswrapper[4979]: I1204 12:03:29.234844 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e002b60f895518acc573949449b33f96e4d7b791f83fbbe60399e9dd7784afa0" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.694824 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x2hnf" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" probeResult="failure" output=< Dec 04 12:03:32 crc kubenswrapper[4979]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 12:03:32 crc kubenswrapper[4979]: > Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.702062 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.712476 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.927597 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x2hnf-config-bjqlp"] Dec 04 12:03:32 crc kubenswrapper[4979]: E1204 12:03:32.928284 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e14c1da3-5204-4d84-94fd-039b1cef8af3" containerName="swift-ring-rebalance" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.928387 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e14c1da3-5204-4d84-94fd-039b1cef8af3" containerName="swift-ring-rebalance" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.928794 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e14c1da3-5204-4d84-94fd-039b1cef8af3" containerName="swift-ring-rebalance" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.929397 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.934708 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.956534 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x2hnf-config-bjqlp"] Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.981803 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-scripts\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.981858 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-additional-scripts\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.981907 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ftzv\" (UniqueName: \"kubernetes.io/projected/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-kube-api-access-9ftzv\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.981961 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run-ovn\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.981985 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-log-ovn\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:32 crc kubenswrapper[4979]: I1204 12:03:32.982003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.082986 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083062 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-scripts\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083117 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-additional-scripts\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083178 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ftzv\" (UniqueName: \"kubernetes.io/projected/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-kube-api-access-9ftzv\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083261 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run-ovn\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083317 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-log-ovn\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083663 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-log-ovn\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.083720 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.084453 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run-ovn\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.085128 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-additional-scripts\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.087082 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-scripts\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.115587 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ftzv\" (UniqueName: \"kubernetes.io/projected/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-kube-api-access-9ftzv\") pod \"ovn-controller-x2hnf-config-bjqlp\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.303379 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:33 crc kubenswrapper[4979]: I1204 12:03:33.824324 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x2hnf-config-bjqlp"] Dec 04 12:03:33 crc kubenswrapper[4979]: W1204 12:03:33.828006 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f8582b7_4f7f_4cec_b9f9_d01dafcf8393.slice/crio-553bc99b80adeb19017534c27928f3387dfff512a38fb8260794121ca0ca3888 WatchSource:0}: Error finding container 553bc99b80adeb19017534c27928f3387dfff512a38fb8260794121ca0ca3888: Status 404 returned error can't find the container with id 553bc99b80adeb19017534c27928f3387dfff512a38fb8260794121ca0ca3888 Dec 04 12:03:34 crc kubenswrapper[4979]: I1204 12:03:34.275109 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-bjqlp" event={"ID":"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393","Type":"ContainerStarted","Data":"553bc99b80adeb19017534c27928f3387dfff512a38fb8260794121ca0ca3888"} Dec 04 12:03:35 crc kubenswrapper[4979]: I1204 12:03:35.283976 4979 generic.go:334] "Generic (PLEG): container finished" podID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerID="ddc7fe209b64e40b2ff31581c3560cda53976230e0b9d95e54824b9db34296d8" exitCode=0 Dec 04 12:03:35 crc kubenswrapper[4979]: I1204 12:03:35.284051 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0fad7195-d19a-48ce-ad3d-4c67b53d2974","Type":"ContainerDied","Data":"ddc7fe209b64e40b2ff31581c3560cda53976230e0b9d95e54824b9db34296d8"} Dec 04 12:03:35 crc kubenswrapper[4979]: I1204 12:03:35.285837 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-bjqlp" event={"ID":"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393","Type":"ContainerStarted","Data":"abe4d0cf444f7acbf247979eb35483dd200bf8268c003927ff768849ca0addcc"} Dec 04 12:03:35 crc kubenswrapper[4979]: I1204 12:03:35.326075 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-x2hnf-config-bjqlp" podStartSLOduration=3.326043622 podStartE2EDuration="3.326043622s" podCreationTimestamp="2025-12-04 12:03:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:35.320742238 +0000 UTC m=+1239.595038042" watchObservedRunningTime="2025-12-04 12:03:35.326043622 +0000 UTC m=+1239.600339416" Dec 04 12:03:36 crc kubenswrapper[4979]: I1204 12:03:36.298681 4979 generic.go:334] "Generic (PLEG): container finished" podID="6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" containerID="abe4d0cf444f7acbf247979eb35483dd200bf8268c003927ff768849ca0addcc" exitCode=0 Dec 04 12:03:36 crc kubenswrapper[4979]: I1204 12:03:36.298731 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-bjqlp" event={"ID":"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393","Type":"ContainerDied","Data":"abe4d0cf444f7acbf247979eb35483dd200bf8268c003927ff768849ca0addcc"} Dec 04 12:03:37 crc kubenswrapper[4979]: I1204 12:03:37.694902 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-x2hnf" Dec 04 12:03:42 crc kubenswrapper[4979]: I1204 12:03:42.318868 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:42 crc kubenswrapper[4979]: I1204 12:03:42.325413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"swift-storage-0\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " pod="openstack/swift-storage-0" Dec 04 12:03:42 crc kubenswrapper[4979]: I1204 12:03:42.354205 4979 generic.go:334] "Generic (PLEG): container finished" podID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerID="23f122939761be5f1b8c15a337a24423951f7eb2b8b6183e7f17f9e0cc30c1e6" exitCode=0 Dec 04 12:03:42 crc kubenswrapper[4979]: I1204 12:03:42.354263 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48","Type":"ContainerDied","Data":"23f122939761be5f1b8c15a337a24423951f7eb2b8b6183e7f17f9e0cc30c1e6"} Dec 04 12:03:42 crc kubenswrapper[4979]: I1204 12:03:42.506977 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 12:03:49 crc kubenswrapper[4979]: E1204 12:03:49.536362 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Dec 04 12:03:49 crc kubenswrapper[4979]: E1204 12:03:49.538404 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h4849,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-8kdt5_openstack(f4f85b75-ee47-4c71-822c-e30cd4f3492c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:03:49 crc kubenswrapper[4979]: E1204 12:03:49.539928 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-8kdt5" podUID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.730014 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.916891 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run\") pod \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.917183 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9ftzv\" (UniqueName: \"kubernetes.io/projected/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-kube-api-access-9ftzv\") pod \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.917270 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-additional-scripts\") pod \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.917320 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-log-ovn\") pod \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.917340 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run-ovn\") pod \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.917408 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-scripts\") pod \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\" (UID: \"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393\") " Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.917087 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run" (OuterVolumeSpecName: "var-run") pod "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" (UID: "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.918593 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" (UID: "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.918645 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" (UID: "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.918656 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-scripts" (OuterVolumeSpecName: "scripts") pod "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" (UID: "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.919459 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" (UID: "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:49 crc kubenswrapper[4979]: I1204 12:03:49.923033 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-kube-api-access-9ftzv" (OuterVolumeSpecName: "kube-api-access-9ftzv") pod "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" (UID: "6f8582b7-4f7f-4cec-b9f9-d01dafcf8393"). InnerVolumeSpecName "kube-api-access-9ftzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.019879 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.020127 4979 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.020169 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9ftzv\" (UniqueName: \"kubernetes.io/projected/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-kube-api-access-9ftzv\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.020193 4979 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.020212 4979 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.020229 4979 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.133471 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.426033 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-bjqlp" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.426478 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-bjqlp" event={"ID":"6f8582b7-4f7f-4cec-b9f9-d01dafcf8393","Type":"ContainerDied","Data":"553bc99b80adeb19017534c27928f3387dfff512a38fb8260794121ca0ca3888"} Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.426528 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="553bc99b80adeb19017534c27928f3387dfff512a38fb8260794121ca0ca3888" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.429281 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0fad7195-d19a-48ce-ad3d-4c67b53d2974","Type":"ContainerStarted","Data":"a170026606f642d3fddc6953a178bf2cc7cbe7322dc226b38b4e679b7df1e228"} Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.429617 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.433155 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48","Type":"ContainerStarted","Data":"ea56050492cab7388c6ef7c5a9645f7cda3ba4238e27b00f2097646a63874d12"} Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.433377 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.435605 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"2760da9775e0b47f8dd8efebeda8de2754523160f4600c33f3a61dec8e1bccaf"} Dec 04 12:03:50 crc kubenswrapper[4979]: E1204 12:03:50.437053 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-8kdt5" podUID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.461939 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=50.800150023 podStartE2EDuration="1m27.461912015s" podCreationTimestamp="2025-12-04 12:02:23 +0000 UTC" firstStartedPulling="2025-12-04 12:02:25.251500364 +0000 UTC m=+1169.525796168" lastFinishedPulling="2025-12-04 12:03:01.913262336 +0000 UTC m=+1206.187558160" observedRunningTime="2025-12-04 12:03:50.45583227 +0000 UTC m=+1254.730128084" watchObservedRunningTime="2025-12-04 12:03:50.461912015 +0000 UTC m=+1254.736207849" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.517592 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=-9223371948.33721 podStartE2EDuration="1m28.517565207s" podCreationTimestamp="2025-12-04 12:02:22 +0000 UTC" firstStartedPulling="2025-12-04 12:02:25.182596131 +0000 UTC m=+1169.456891935" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:03:50.514633707 +0000 UTC m=+1254.788929511" watchObservedRunningTime="2025-12-04 12:03:50.517565207 +0000 UTC m=+1254.791861031" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.827292 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x2hnf-config-bjqlp"] Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.843559 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-x2hnf-config-bjqlp"] Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.922524 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-x2hnf-config-7mnmx"] Dec 04 12:03:50 crc kubenswrapper[4979]: E1204 12:03:50.922911 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" containerName="ovn-config" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.922929 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" containerName="ovn-config" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.923141 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" containerName="ovn-config" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.923776 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.926237 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 12:03:50 crc kubenswrapper[4979]: I1204 12:03:50.935420 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x2hnf-config-7mnmx"] Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.035236 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccgmn\" (UniqueName: \"kubernetes.io/projected/c7ef265d-0acd-4c89-937c-fc9c56853b49-kube-api-access-ccgmn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.035288 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run-ovn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.035338 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.035419 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-log-ovn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.035460 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-additional-scripts\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.035507 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-scripts\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137120 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccgmn\" (UniqueName: \"kubernetes.io/projected/c7ef265d-0acd-4c89-937c-fc9c56853b49-kube-api-access-ccgmn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137270 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run-ovn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137326 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137392 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-log-ovn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137422 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-additional-scripts\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137488 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-scripts\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.137935 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.138079 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-log-ovn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.138285 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run-ovn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.138653 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-additional-scripts\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.139518 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-scripts\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.178040 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccgmn\" (UniqueName: \"kubernetes.io/projected/c7ef265d-0acd-4c89-937c-fc9c56853b49-kube-api-access-ccgmn\") pod \"ovn-controller-x2hnf-config-7mnmx\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.244912 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:51 crc kubenswrapper[4979]: I1204 12:03:51.542457 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-x2hnf-config-7mnmx"] Dec 04 12:03:51 crc kubenswrapper[4979]: W1204 12:03:51.544755 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7ef265d_0acd_4c89_937c_fc9c56853b49.slice/crio-108921ce196c08fc174ad4958a7c9441113eef58456944f59dc95fcc4f53eec0 WatchSource:0}: Error finding container 108921ce196c08fc174ad4958a7c9441113eef58456944f59dc95fcc4f53eec0: Status 404 returned error can't find the container with id 108921ce196c08fc174ad4958a7c9441113eef58456944f59dc95fcc4f53eec0 Dec 04 12:03:52 crc kubenswrapper[4979]: I1204 12:03:52.208992 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f8582b7-4f7f-4cec-b9f9-d01dafcf8393" path="/var/lib/kubelet/pods/6f8582b7-4f7f-4cec-b9f9-d01dafcf8393/volumes" Dec 04 12:03:52 crc kubenswrapper[4979]: I1204 12:03:52.457658 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6"} Dec 04 12:03:52 crc kubenswrapper[4979]: I1204 12:03:52.457705 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254"} Dec 04 12:03:52 crc kubenswrapper[4979]: I1204 12:03:52.462333 4979 generic.go:334] "Generic (PLEG): container finished" podID="c7ef265d-0acd-4c89-937c-fc9c56853b49" containerID="84ca39ca4cf3463b208bb9394e262a318f737de5230f3c751a488c39ef49a00f" exitCode=0 Dec 04 12:03:52 crc kubenswrapper[4979]: I1204 12:03:52.462368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-7mnmx" event={"ID":"c7ef265d-0acd-4c89-937c-fc9c56853b49","Type":"ContainerDied","Data":"84ca39ca4cf3463b208bb9394e262a318f737de5230f3c751a488c39ef49a00f"} Dec 04 12:03:52 crc kubenswrapper[4979]: I1204 12:03:52.462388 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-7mnmx" event={"ID":"c7ef265d-0acd-4c89-937c-fc9c56853b49","Type":"ContainerStarted","Data":"108921ce196c08fc174ad4958a7c9441113eef58456944f59dc95fcc4f53eec0"} Dec 04 12:03:53 crc kubenswrapper[4979]: I1204 12:03:53.478139 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15"} Dec 04 12:03:53 crc kubenswrapper[4979]: I1204 12:03:53.478515 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238"} Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.347855 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.488666 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf-config-7mnmx" event={"ID":"c7ef265d-0acd-4c89-937c-fc9c56853b49","Type":"ContainerDied","Data":"108921ce196c08fc174ad4958a7c9441113eef58456944f59dc95fcc4f53eec0"} Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.488716 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="108921ce196c08fc174ad4958a7c9441113eef58456944f59dc95fcc4f53eec0" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.488746 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf-config-7mnmx" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.511891 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-log-ovn\") pod \"c7ef265d-0acd-4c89-937c-fc9c56853b49\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.511942 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccgmn\" (UniqueName: \"kubernetes.io/projected/c7ef265d-0acd-4c89-937c-fc9c56853b49-kube-api-access-ccgmn\") pod \"c7ef265d-0acd-4c89-937c-fc9c56853b49\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.512097 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-scripts\") pod \"c7ef265d-0acd-4c89-937c-fc9c56853b49\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.512121 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run\") pod \"c7ef265d-0acd-4c89-937c-fc9c56853b49\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.512156 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run-ovn\") pod \"c7ef265d-0acd-4c89-937c-fc9c56853b49\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.512180 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-additional-scripts\") pod \"c7ef265d-0acd-4c89-937c-fc9c56853b49\" (UID: \"c7ef265d-0acd-4c89-937c-fc9c56853b49\") " Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.512990 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c7ef265d-0acd-4c89-937c-fc9c56853b49" (UID: "c7ef265d-0acd-4c89-937c-fc9c56853b49"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.513508 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c7ef265d-0acd-4c89-937c-fc9c56853b49" (UID: "c7ef265d-0acd-4c89-937c-fc9c56853b49"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.513558 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run" (OuterVolumeSpecName: "var-run") pod "c7ef265d-0acd-4c89-937c-fc9c56853b49" (UID: "c7ef265d-0acd-4c89-937c-fc9c56853b49"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.514596 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-scripts" (OuterVolumeSpecName: "scripts") pod "c7ef265d-0acd-4c89-937c-fc9c56853b49" (UID: "c7ef265d-0acd-4c89-937c-fc9c56853b49"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.514640 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c7ef265d-0acd-4c89-937c-fc9c56853b49" (UID: "c7ef265d-0acd-4c89-937c-fc9c56853b49"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.521484 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7ef265d-0acd-4c89-937c-fc9c56853b49-kube-api-access-ccgmn" (OuterVolumeSpecName: "kube-api-access-ccgmn") pod "c7ef265d-0acd-4c89-937c-fc9c56853b49" (UID: "c7ef265d-0acd-4c89-937c-fc9c56853b49"). InnerVolumeSpecName "kube-api-access-ccgmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.613868 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.613906 4979 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.613918 4979 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.613927 4979 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c7ef265d-0acd-4c89-937c-fc9c56853b49-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.613938 4979 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c7ef265d-0acd-4c89-937c-fc9c56853b49-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:54 crc kubenswrapper[4979]: I1204 12:03:54.613948 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccgmn\" (UniqueName: \"kubernetes.io/projected/c7ef265d-0acd-4c89-937c-fc9c56853b49-kube-api-access-ccgmn\") on node \"crc\" DevicePath \"\"" Dec 04 12:03:55 crc kubenswrapper[4979]: I1204 12:03:55.465266 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x2hnf-config-7mnmx"] Dec 04 12:03:55 crc kubenswrapper[4979]: I1204 12:03:55.475858 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-x2hnf-config-7mnmx"] Dec 04 12:03:55 crc kubenswrapper[4979]: I1204 12:03:55.500570 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01"} Dec 04 12:03:55 crc kubenswrapper[4979]: I1204 12:03:55.501072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757"} Dec 04 12:03:55 crc kubenswrapper[4979]: I1204 12:03:55.501143 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf"} Dec 04 12:03:55 crc kubenswrapper[4979]: I1204 12:03:55.501212 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946"} Dec 04 12:03:56 crc kubenswrapper[4979]: I1204 12:03:56.217087 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7ef265d-0acd-4c89-937c-fc9c56853b49" path="/var/lib/kubelet/pods/c7ef265d-0acd-4c89-937c-fc9c56853b49/volumes" Dec 04 12:03:56 crc kubenswrapper[4979]: I1204 12:03:56.522337 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5"} Dec 04 12:03:56 crc kubenswrapper[4979]: I1204 12:03:56.522383 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e"} Dec 04 12:03:56 crc kubenswrapper[4979]: I1204 12:03:56.522397 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab"} Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.550539 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe"} Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.550844 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509"} Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.550870 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957"} Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.550881 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerStarted","Data":"edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55"} Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.591762 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=42.787685196 podStartE2EDuration="48.591740152s" podCreationTimestamp="2025-12-04 12:03:09 +0000 UTC" firstStartedPulling="2025-12-04 12:03:50.144937493 +0000 UTC m=+1254.419233297" lastFinishedPulling="2025-12-04 12:03:55.948992439 +0000 UTC m=+1260.223288253" observedRunningTime="2025-12-04 12:03:57.585289208 +0000 UTC m=+1261.859585032" watchObservedRunningTime="2025-12-04 12:03:57.591740152 +0000 UTC m=+1261.866035956" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.865385 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rmf45"] Dec 04 12:03:57 crc kubenswrapper[4979]: E1204 12:03:57.867015 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7ef265d-0acd-4c89-937c-fc9c56853b49" containerName="ovn-config" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.867064 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7ef265d-0acd-4c89-937c-fc9c56853b49" containerName="ovn-config" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.867409 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7ef265d-0acd-4c89-937c-fc9c56853b49" containerName="ovn-config" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.870478 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.872483 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.878729 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rmf45"] Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.969933 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.969978 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.970099 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.970148 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmrq2\" (UniqueName: \"kubernetes.io/projected/333d2328-de05-4d66-a8b1-d0585ddf5f32-kube-api-access-dmrq2\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.970188 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:57 crc kubenswrapper[4979]: I1204 12:03:57.970392 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-config\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.071604 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.072002 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.072260 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.072284 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmrq2\" (UniqueName: \"kubernetes.io/projected/333d2328-de05-4d66-a8b1-d0585ddf5f32-kube-api-access-dmrq2\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.072332 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.072418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-config\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.072807 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.073526 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-config\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.073583 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.074317 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.074611 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.099818 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmrq2\" (UniqueName: \"kubernetes.io/projected/333d2328-de05-4d66-a8b1-d0585ddf5f32-kube-api-access-dmrq2\") pod \"dnsmasq-dns-764c5664d7-rmf45\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.186980 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:03:58 crc kubenswrapper[4979]: I1204 12:03:58.656835 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rmf45"] Dec 04 12:03:59 crc kubenswrapper[4979]: I1204 12:03:59.569937 4979 generic.go:334] "Generic (PLEG): container finished" podID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerID="194f4362236a1ef75fb2d02b607a54c1b599af7c069ec7c864abb20274f24377" exitCode=0 Dec 04 12:03:59 crc kubenswrapper[4979]: I1204 12:03:59.570152 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" event={"ID":"333d2328-de05-4d66-a8b1-d0585ddf5f32","Type":"ContainerDied","Data":"194f4362236a1ef75fb2d02b607a54c1b599af7c069ec7c864abb20274f24377"} Dec 04 12:03:59 crc kubenswrapper[4979]: I1204 12:03:59.570234 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" event={"ID":"333d2328-de05-4d66-a8b1-d0585ddf5f32","Type":"ContainerStarted","Data":"a6d55e841376361a09ee4675cf285d5ee18832620f9e4e42f8359315c118f705"} Dec 04 12:04:00 crc kubenswrapper[4979]: I1204 12:04:00.581495 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" event={"ID":"333d2328-de05-4d66-a8b1-d0585ddf5f32","Type":"ContainerStarted","Data":"c8be71aa2ab63e7bba5003b34e684ff345e36633a9b0398fbf6538afd399a7a6"} Dec 04 12:04:00 crc kubenswrapper[4979]: I1204 12:04:00.581871 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:04:00 crc kubenswrapper[4979]: I1204 12:04:00.602182 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" podStartSLOduration=3.602162845 podStartE2EDuration="3.602162845s" podCreationTimestamp="2025-12-04 12:03:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:00.597084658 +0000 UTC m=+1264.871380472" watchObservedRunningTime="2025-12-04 12:04:00.602162845 +0000 UTC m=+1264.876458659" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.590613 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.621229 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8kdt5" event={"ID":"f4f85b75-ee47-4c71-822c-e30cd4f3492c","Type":"ContainerStarted","Data":"3c64b869abf2343d534fe0e3db701fe908e10cca6fa7036f2bb28093bd4e5f7e"} Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.631402 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.654431 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-8kdt5" podStartSLOduration=1.899592465 podStartE2EDuration="41.654417915s" podCreationTimestamp="2025-12-04 12:03:23 +0000 UTC" firstStartedPulling="2025-12-04 12:03:24.038249603 +0000 UTC m=+1228.312545417" lastFinishedPulling="2025-12-04 12:04:03.793075043 +0000 UTC m=+1268.067370867" observedRunningTime="2025-12-04 12:04:04.65126721 +0000 UTC m=+1268.925563014" watchObservedRunningTime="2025-12-04 12:04:04.654417915 +0000 UTC m=+1268.928713719" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.979208 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-znf8s"] Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.980858 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.988156 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-ee53-account-create-update-qlklb"] Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.989389 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.991478 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 12:04:04 crc kubenswrapper[4979]: I1204 12:04:04.999798 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-znf8s"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.009152 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ee53-account-create-update-qlklb"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.094486 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-swzf4"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.096575 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.106002 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-operator-scripts\") pod \"barbican-db-create-znf8s\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.106102 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcxxk\" (UniqueName: \"kubernetes.io/projected/ad7d638c-e396-403e-974b-864194d22469-kube-api-access-wcxxk\") pod \"barbican-ee53-account-create-update-qlklb\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.106158 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhbpx\" (UniqueName: \"kubernetes.io/projected/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-kube-api-access-xhbpx\") pod \"barbican-db-create-znf8s\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.106193 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7d638c-e396-403e-974b-864194d22469-operator-scripts\") pod \"barbican-ee53-account-create-update-qlklb\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.122761 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-swzf4"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.168781 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-37f5-account-create-update-m9f9n"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.170195 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.175180 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.185789 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-37f5-account-create-update-m9f9n"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.207685 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcxxk\" (UniqueName: \"kubernetes.io/projected/ad7d638c-e396-403e-974b-864194d22469-kube-api-access-wcxxk\") pod \"barbican-ee53-account-create-update-qlklb\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.207736 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7eb59f9-ff28-4260-834c-dffefdc7db5e-operator-scripts\") pod \"cinder-db-create-swzf4\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.207771 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxhfp\" (UniqueName: \"kubernetes.io/projected/a7eb59f9-ff28-4260-834c-dffefdc7db5e-kube-api-access-cxhfp\") pod \"cinder-db-create-swzf4\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.207803 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhbpx\" (UniqueName: \"kubernetes.io/projected/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-kube-api-access-xhbpx\") pod \"barbican-db-create-znf8s\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.207875 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7d638c-e396-403e-974b-864194d22469-operator-scripts\") pod \"barbican-ee53-account-create-update-qlklb\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.207937 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-operator-scripts\") pod \"barbican-db-create-znf8s\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.208665 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-operator-scripts\") pod \"barbican-db-create-znf8s\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.209524 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7d638c-e396-403e-974b-864194d22469-operator-scripts\") pod \"barbican-ee53-account-create-update-qlklb\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.235225 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhbpx\" (UniqueName: \"kubernetes.io/projected/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-kube-api-access-xhbpx\") pod \"barbican-db-create-znf8s\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.245796 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcxxk\" (UniqueName: \"kubernetes.io/projected/ad7d638c-e396-403e-974b-864194d22469-kube-api-access-wcxxk\") pod \"barbican-ee53-account-create-update-qlklb\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.245864 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-4q48m"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.246847 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.261167 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.261356 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zbvzn" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.261451 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.261917 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.265683 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4q48m"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.301998 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.310002 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/173171c8-4339-42ae-a4bc-be3cdc3420be-operator-scripts\") pod \"cinder-37f5-account-create-update-m9f9n\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.310082 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7eb59f9-ff28-4260-834c-dffefdc7db5e-operator-scripts\") pod \"cinder-db-create-swzf4\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.310114 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts2nc\" (UniqueName: \"kubernetes.io/projected/173171c8-4339-42ae-a4bc-be3cdc3420be-kube-api-access-ts2nc\") pod \"cinder-37f5-account-create-update-m9f9n\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.310144 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxhfp\" (UniqueName: \"kubernetes.io/projected/a7eb59f9-ff28-4260-834c-dffefdc7db5e-kube-api-access-cxhfp\") pod \"cinder-db-create-swzf4\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.312613 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.313661 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7eb59f9-ff28-4260-834c-dffefdc7db5e-operator-scripts\") pod \"cinder-db-create-swzf4\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.334288 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxhfp\" (UniqueName: \"kubernetes.io/projected/a7eb59f9-ff28-4260-834c-dffefdc7db5e-kube-api-access-cxhfp\") pod \"cinder-db-create-swzf4\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.379658 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-47fe-account-create-update-dpg62"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.381138 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.387266 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.395839 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-47fe-account-create-update-dpg62"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.411638 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8cqc\" (UniqueName: \"kubernetes.io/projected/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-kube-api-access-f8cqc\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.411686 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-combined-ca-bundle\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.411752 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-config-data\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.411827 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/173171c8-4339-42ae-a4bc-be3cdc3420be-operator-scripts\") pod \"cinder-37f5-account-create-update-m9f9n\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.411862 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts2nc\" (UniqueName: \"kubernetes.io/projected/173171c8-4339-42ae-a4bc-be3cdc3420be-kube-api-access-ts2nc\") pod \"cinder-37f5-account-create-update-m9f9n\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.413088 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/173171c8-4339-42ae-a4bc-be3cdc3420be-operator-scripts\") pod \"cinder-37f5-account-create-update-m9f9n\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.425338 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.433404 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts2nc\" (UniqueName: \"kubernetes.io/projected/173171c8-4339-42ae-a4bc-be3cdc3420be-kube-api-access-ts2nc\") pod \"cinder-37f5-account-create-update-m9f9n\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.486340 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-b4thg"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.487774 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.489101 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.507815 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b4thg"] Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.515218 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dxmr\" (UniqueName: \"kubernetes.io/projected/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-kube-api-access-7dxmr\") pod \"neutron-47fe-account-create-update-dpg62\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.515329 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8cqc\" (UniqueName: \"kubernetes.io/projected/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-kube-api-access-f8cqc\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.515351 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-combined-ca-bundle\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.515396 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-operator-scripts\") pod \"neutron-47fe-account-create-update-dpg62\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.515415 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-config-data\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.523655 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-config-data\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.529984 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-combined-ca-bundle\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.548982 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8cqc\" (UniqueName: \"kubernetes.io/projected/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-kube-api-access-f8cqc\") pod \"keystone-db-sync-4q48m\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.619747 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-operator-scripts\") pod \"neutron-47fe-account-create-update-dpg62\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.620281 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1177173d-34cc-4f8c-9ac9-e2661449305f-operator-scripts\") pod \"neutron-db-create-b4thg\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.620380 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dxmr\" (UniqueName: \"kubernetes.io/projected/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-kube-api-access-7dxmr\") pod \"neutron-47fe-account-create-update-dpg62\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.620444 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nxxq\" (UniqueName: \"kubernetes.io/projected/1177173d-34cc-4f8c-9ac9-e2661449305f-kube-api-access-9nxxq\") pod \"neutron-db-create-b4thg\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.621181 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-operator-scripts\") pod \"neutron-47fe-account-create-update-dpg62\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.650850 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dxmr\" (UniqueName: \"kubernetes.io/projected/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-kube-api-access-7dxmr\") pod \"neutron-47fe-account-create-update-dpg62\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.662104 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ee53-account-create-update-qlklb"] Dec 04 12:04:05 crc kubenswrapper[4979]: W1204 12:04:05.669580 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad7d638c_e396_403e_974b_864194d22469.slice/crio-2498dc0f89fd873f2162c184951f76c42b5bb11c85460d804800f4020dbc5d67 WatchSource:0}: Error finding container 2498dc0f89fd873f2162c184951f76c42b5bb11c85460d804800f4020dbc5d67: Status 404 returned error can't find the container with id 2498dc0f89fd873f2162c184951f76c42b5bb11c85460d804800f4020dbc5d67 Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.720136 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.721700 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nxxq\" (UniqueName: \"kubernetes.io/projected/1177173d-34cc-4f8c-9ac9-e2661449305f-kube-api-access-9nxxq\") pod \"neutron-db-create-b4thg\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.721847 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1177173d-34cc-4f8c-9ac9-e2661449305f-operator-scripts\") pod \"neutron-db-create-b4thg\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.722629 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1177173d-34cc-4f8c-9ac9-e2661449305f-operator-scripts\") pod \"neutron-db-create-b4thg\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:05 crc kubenswrapper[4979]: I1204 12:04:05.727821 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:05.741930 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nxxq\" (UniqueName: \"kubernetes.io/projected/1177173d-34cc-4f8c-9ac9-e2661449305f-kube-api-access-9nxxq\") pod \"neutron-db-create-b4thg\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:05.814151 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-37f5-account-create-update-m9f9n"] Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:05.824239 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:06 crc kubenswrapper[4979]: W1204 12:04:05.844526 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod173171c8_4339_42ae_a4bc_be3cdc3420be.slice/crio-7c4080d7c8c1afa345a935b3a4835b168b11a2a295aae0c01fdb0c23529f1f09 WatchSource:0}: Error finding container 7c4080d7c8c1afa345a935b3a4835b168b11a2a295aae0c01fdb0c23529f1f09: Status 404 returned error can't find the container with id 7c4080d7c8c1afa345a935b3a4835b168b11a2a295aae0c01fdb0c23529f1f09 Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:05.937157 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-znf8s"] Dec 04 12:04:06 crc kubenswrapper[4979]: W1204 12:04:05.949472 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode88366f6_4404_46ae_8538_a7bd3cc9bbb5.slice/crio-36085710ec82d4db1a001b4049cf2b01a6b15f0b68599a4a82d5ae3c71100479 WatchSource:0}: Error finding container 36085710ec82d4db1a001b4049cf2b01a6b15f0b68599a4a82d5ae3c71100479: Status 404 returned error can't find the container with id 36085710ec82d4db1a001b4049cf2b01a6b15f0b68599a4a82d5ae3c71100479 Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:05.972834 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-swzf4"] Dec 04 12:04:06 crc kubenswrapper[4979]: W1204 12:04:05.983234 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7eb59f9_ff28_4260_834c_dffefdc7db5e.slice/crio-b9d78c1c6988255cca42b6a6484f5520c2b0b95ca58c79fb776eb579150357d0 WatchSource:0}: Error finding container b9d78c1c6988255cca42b6a6484f5520c2b0b95ca58c79fb776eb579150357d0: Status 404 returned error can't find the container with id b9d78c1c6988255cca42b6a6484f5520c2b0b95ca58c79fb776eb579150357d0 Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.661255 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-swzf4" event={"ID":"a7eb59f9-ff28-4260-834c-dffefdc7db5e","Type":"ContainerStarted","Data":"a386179a59dd07a6713af9202101515d8f0def69fccedf342f0246871f55cb4b"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.661639 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-swzf4" event={"ID":"a7eb59f9-ff28-4260-834c-dffefdc7db5e","Type":"ContainerStarted","Data":"b9d78c1c6988255cca42b6a6484f5520c2b0b95ca58c79fb776eb579150357d0"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.665151 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-37f5-account-create-update-m9f9n" event={"ID":"173171c8-4339-42ae-a4bc-be3cdc3420be","Type":"ContainerStarted","Data":"79e2c3ff30be15bd922b3998cf47243382b6d384d3d6b9b8c6698361d07eb8a1"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.665182 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-37f5-account-create-update-m9f9n" event={"ID":"173171c8-4339-42ae-a4bc-be3cdc3420be","Type":"ContainerStarted","Data":"7c4080d7c8c1afa345a935b3a4835b168b11a2a295aae0c01fdb0c23529f1f09"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.666910 4979 generic.go:334] "Generic (PLEG): container finished" podID="ad7d638c-e396-403e-974b-864194d22469" containerID="3854a6ae681a4089c468e8e386b10156644e06f17c8fc68fe7422dbcc59dbb35" exitCode=0 Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.666957 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ee53-account-create-update-qlklb" event={"ID":"ad7d638c-e396-403e-974b-864194d22469","Type":"ContainerDied","Data":"3854a6ae681a4089c468e8e386b10156644e06f17c8fc68fe7422dbcc59dbb35"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.666975 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ee53-account-create-update-qlklb" event={"ID":"ad7d638c-e396-403e-974b-864194d22469","Type":"ContainerStarted","Data":"2498dc0f89fd873f2162c184951f76c42b5bb11c85460d804800f4020dbc5d67"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.668942 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-znf8s" event={"ID":"e88366f6-4404-46ae-8538-a7bd3cc9bbb5","Type":"ContainerStarted","Data":"fc52fb357d41227cb0272a57c0f91beb6e46c9edf9d85b28eaeda62c1c8e1dfe"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.668975 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-znf8s" event={"ID":"e88366f6-4404-46ae-8538-a7bd3cc9bbb5","Type":"ContainerStarted","Data":"36085710ec82d4db1a001b4049cf2b01a6b15f0b68599a4a82d5ae3c71100479"} Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.674933 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-swzf4" podStartSLOduration=1.674916273 podStartE2EDuration="1.674916273s" podCreationTimestamp="2025-12-04 12:04:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:06.674158202 +0000 UTC m=+1270.948454006" watchObservedRunningTime="2025-12-04 12:04:06.674916273 +0000 UTC m=+1270.949212077" Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.702136 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-znf8s" podStartSLOduration=2.702112932 podStartE2EDuration="2.702112932s" podCreationTimestamp="2025-12-04 12:04:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:06.694839624 +0000 UTC m=+1270.969135418" watchObservedRunningTime="2025-12-04 12:04:06.702112932 +0000 UTC m=+1270.976408736" Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.742287 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-37f5-account-create-update-m9f9n" podStartSLOduration=1.742270683 podStartE2EDuration="1.742270683s" podCreationTimestamp="2025-12-04 12:04:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:06.732610271 +0000 UTC m=+1271.006906075" watchObservedRunningTime="2025-12-04 12:04:06.742270683 +0000 UTC m=+1271.016566487" Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.949610 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-47fe-account-create-update-dpg62"] Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.958399 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-b4thg"] Dec 04 12:04:06 crc kubenswrapper[4979]: W1204 12:04:06.963569 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4dd2a291_3cb5_43ab_8f6c_c7b9517f3455.slice/crio-38c0f184e5d4b0adbcf0ac12901708db3b7fa08d2231dc3b7a47fc1ee5983b9e WatchSource:0}: Error finding container 38c0f184e5d4b0adbcf0ac12901708db3b7fa08d2231dc3b7a47fc1ee5983b9e: Status 404 returned error can't find the container with id 38c0f184e5d4b0adbcf0ac12901708db3b7fa08d2231dc3b7a47fc1ee5983b9e Dec 04 12:04:06 crc kubenswrapper[4979]: W1204 12:04:06.964749 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1177173d_34cc_4f8c_9ac9_e2661449305f.slice/crio-7104c25979bd26ec6369c681027cd71cb5bc082d138e613bc650dc1276171d80 WatchSource:0}: Error finding container 7104c25979bd26ec6369c681027cd71cb5bc082d138e613bc650dc1276171d80: Status 404 returned error can't find the container with id 7104c25979bd26ec6369c681027cd71cb5bc082d138e613bc650dc1276171d80 Dec 04 12:04:06 crc kubenswrapper[4979]: I1204 12:04:06.976325 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-4q48m"] Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.696054 4979 generic.go:334] "Generic (PLEG): container finished" podID="173171c8-4339-42ae-a4bc-be3cdc3420be" containerID="79e2c3ff30be15bd922b3998cf47243382b6d384d3d6b9b8c6698361d07eb8a1" exitCode=0 Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.696256 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-37f5-account-create-update-m9f9n" event={"ID":"173171c8-4339-42ae-a4bc-be3cdc3420be","Type":"ContainerDied","Data":"79e2c3ff30be15bd922b3998cf47243382b6d384d3d6b9b8c6698361d07eb8a1"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.699814 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4q48m" event={"ID":"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b","Type":"ContainerStarted","Data":"e039cb4eb68ae96961c1acf7d288c4fb5927cecd86df9e89a69b6a829f1e2a3a"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.702727 4979 generic.go:334] "Generic (PLEG): container finished" podID="1177173d-34cc-4f8c-9ac9-e2661449305f" containerID="e2e6214d14320ff207b7aebe8cd3428ad03cff707cf31a646cb28e7394285d50" exitCode=0 Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.702843 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b4thg" event={"ID":"1177173d-34cc-4f8c-9ac9-e2661449305f","Type":"ContainerDied","Data":"e2e6214d14320ff207b7aebe8cd3428ad03cff707cf31a646cb28e7394285d50"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.702881 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b4thg" event={"ID":"1177173d-34cc-4f8c-9ac9-e2661449305f","Type":"ContainerStarted","Data":"7104c25979bd26ec6369c681027cd71cb5bc082d138e613bc650dc1276171d80"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.709561 4979 generic.go:334] "Generic (PLEG): container finished" podID="e88366f6-4404-46ae-8538-a7bd3cc9bbb5" containerID="fc52fb357d41227cb0272a57c0f91beb6e46c9edf9d85b28eaeda62c1c8e1dfe" exitCode=0 Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.709603 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-znf8s" event={"ID":"e88366f6-4404-46ae-8538-a7bd3cc9bbb5","Type":"ContainerDied","Data":"fc52fb357d41227cb0272a57c0f91beb6e46c9edf9d85b28eaeda62c1c8e1dfe"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.710667 4979 generic.go:334] "Generic (PLEG): container finished" podID="4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" containerID="ddb76e21454d411642ff491c233f19bf9c4483d4330deff77afae441d539dc77" exitCode=0 Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.710705 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-47fe-account-create-update-dpg62" event={"ID":"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455","Type":"ContainerDied","Data":"ddb76e21454d411642ff491c233f19bf9c4483d4330deff77afae441d539dc77"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.710721 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-47fe-account-create-update-dpg62" event={"ID":"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455","Type":"ContainerStarted","Data":"38c0f184e5d4b0adbcf0ac12901708db3b7fa08d2231dc3b7a47fc1ee5983b9e"} Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.718688 4979 generic.go:334] "Generic (PLEG): container finished" podID="a7eb59f9-ff28-4260-834c-dffefdc7db5e" containerID="a386179a59dd07a6713af9202101515d8f0def69fccedf342f0246871f55cb4b" exitCode=0 Dec 04 12:04:07 crc kubenswrapper[4979]: I1204 12:04:07.718941 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-swzf4" event={"ID":"a7eb59f9-ff28-4260-834c-dffefdc7db5e","Type":"ContainerDied","Data":"a386179a59dd07a6713af9202101515d8f0def69fccedf342f0246871f55cb4b"} Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.017733 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.180018 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcxxk\" (UniqueName: \"kubernetes.io/projected/ad7d638c-e396-403e-974b-864194d22469-kube-api-access-wcxxk\") pod \"ad7d638c-e396-403e-974b-864194d22469\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.180078 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7d638c-e396-403e-974b-864194d22469-operator-scripts\") pod \"ad7d638c-e396-403e-974b-864194d22469\" (UID: \"ad7d638c-e396-403e-974b-864194d22469\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.180792 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad7d638c-e396-403e-974b-864194d22469-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad7d638c-e396-403e-974b-864194d22469" (UID: "ad7d638c-e396-403e-974b-864194d22469"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.184692 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7d638c-e396-403e-974b-864194d22469-kube-api-access-wcxxk" (OuterVolumeSpecName: "kube-api-access-wcxxk") pod "ad7d638c-e396-403e-974b-864194d22469" (UID: "ad7d638c-e396-403e-974b-864194d22469"). InnerVolumeSpecName "kube-api-access-wcxxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.188565 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.251942 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zqq2b"] Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.252895 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-zqq2b" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerName="dnsmasq-dns" containerID="cri-o://3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b" gracePeriod=10 Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.281670 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcxxk\" (UniqueName: \"kubernetes.io/projected/ad7d638c-e396-403e-974b-864194d22469-kube-api-access-wcxxk\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.281695 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad7d638c-e396-403e-974b-864194d22469-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.658866 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.737717 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ee53-account-create-update-qlklb" event={"ID":"ad7d638c-e396-403e-974b-864194d22469","Type":"ContainerDied","Data":"2498dc0f89fd873f2162c184951f76c42b5bb11c85460d804800f4020dbc5d67"} Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.737757 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2498dc0f89fd873f2162c184951f76c42b5bb11c85460d804800f4020dbc5d67" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.737807 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ee53-account-create-update-qlklb" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.746098 4979 generic.go:334] "Generic (PLEG): container finished" podID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerID="3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b" exitCode=0 Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.746236 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zqq2b" event={"ID":"b3c7923a-4a30-4260-ab5a-56d55fbab3e2","Type":"ContainerDied","Data":"3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b"} Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.746283 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-zqq2b" event={"ID":"b3c7923a-4a30-4260-ab5a-56d55fbab3e2","Type":"ContainerDied","Data":"4445780ec18532b9d5a1cad9a33058367cdd4edd9468f5dc0513d4b561b08ebc"} Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.746323 4979 scope.go:117] "RemoveContainer" containerID="3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.746416 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-zqq2b" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.772875 4979 scope.go:117] "RemoveContainer" containerID="8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.792418 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-config\") pod \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.792505 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrj85\" (UniqueName: \"kubernetes.io/projected/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-kube-api-access-xrj85\") pod \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.792540 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-dns-svc\") pod \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.792571 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-nb\") pod \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.792593 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-sb\") pod \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\" (UID: \"b3c7923a-4a30-4260-ab5a-56d55fbab3e2\") " Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.801494 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-kube-api-access-xrj85" (OuterVolumeSpecName: "kube-api-access-xrj85") pod "b3c7923a-4a30-4260-ab5a-56d55fbab3e2" (UID: "b3c7923a-4a30-4260-ab5a-56d55fbab3e2"). InnerVolumeSpecName "kube-api-access-xrj85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.805696 4979 scope.go:117] "RemoveContainer" containerID="3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b" Dec 04 12:04:08 crc kubenswrapper[4979]: E1204 12:04:08.806189 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b\": container with ID starting with 3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b not found: ID does not exist" containerID="3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.806225 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b"} err="failed to get container status \"3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b\": rpc error: code = NotFound desc = could not find container \"3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b\": container with ID starting with 3e43df9d715e76040b7d06202da3936a33d90bf48f0e78de16742aa6819db03b not found: ID does not exist" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.806251 4979 scope.go:117] "RemoveContainer" containerID="8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5" Dec 04 12:04:08 crc kubenswrapper[4979]: E1204 12:04:08.808721 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5\": container with ID starting with 8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5 not found: ID does not exist" containerID="8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.808768 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5"} err="failed to get container status \"8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5\": rpc error: code = NotFound desc = could not find container \"8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5\": container with ID starting with 8e280951407e8dd5483ca5e67cc486f1bb1c09137ec9a84e66b13c878c65f7b5 not found: ID does not exist" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.852105 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-config" (OuterVolumeSpecName: "config") pod "b3c7923a-4a30-4260-ab5a-56d55fbab3e2" (UID: "b3c7923a-4a30-4260-ab5a-56d55fbab3e2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.853118 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b3c7923a-4a30-4260-ab5a-56d55fbab3e2" (UID: "b3c7923a-4a30-4260-ab5a-56d55fbab3e2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.858451 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b3c7923a-4a30-4260-ab5a-56d55fbab3e2" (UID: "b3c7923a-4a30-4260-ab5a-56d55fbab3e2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.861272 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b3c7923a-4a30-4260-ab5a-56d55fbab3e2" (UID: "b3c7923a-4a30-4260-ab5a-56d55fbab3e2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.894037 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrj85\" (UniqueName: \"kubernetes.io/projected/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-kube-api-access-xrj85\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.894074 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.894084 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.894091 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:08 crc kubenswrapper[4979]: I1204 12:04:08.894100 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3c7923a-4a30-4260-ab5a-56d55fbab3e2-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.145757 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.182410 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zqq2b"] Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.192725 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-zqq2b"] Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.299100 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.300247 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts2nc\" (UniqueName: \"kubernetes.io/projected/173171c8-4339-42ae-a4bc-be3cdc3420be-kube-api-access-ts2nc\") pod \"173171c8-4339-42ae-a4bc-be3cdc3420be\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.300327 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/173171c8-4339-42ae-a4bc-be3cdc3420be-operator-scripts\") pod \"173171c8-4339-42ae-a4bc-be3cdc3420be\" (UID: \"173171c8-4339-42ae-a4bc-be3cdc3420be\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.300931 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/173171c8-4339-42ae-a4bc-be3cdc3420be-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "173171c8-4339-42ae-a4bc-be3cdc3420be" (UID: "173171c8-4339-42ae-a4bc-be3cdc3420be"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.303918 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/173171c8-4339-42ae-a4bc-be3cdc3420be-kube-api-access-ts2nc" (OuterVolumeSpecName: "kube-api-access-ts2nc") pod "173171c8-4339-42ae-a4bc-be3cdc3420be" (UID: "173171c8-4339-42ae-a4bc-be3cdc3420be"). InnerVolumeSpecName "kube-api-access-ts2nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.305939 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.321430 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.334951 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.402234 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1177173d-34cc-4f8c-9ac9-e2661449305f-operator-scripts\") pod \"1177173d-34cc-4f8c-9ac9-e2661449305f\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.402481 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhbpx\" (UniqueName: \"kubernetes.io/projected/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-kube-api-access-xhbpx\") pod \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.402589 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-operator-scripts\") pod \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\" (UID: \"e88366f6-4404-46ae-8538-a7bd3cc9bbb5\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.402708 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nxxq\" (UniqueName: \"kubernetes.io/projected/1177173d-34cc-4f8c-9ac9-e2661449305f-kube-api-access-9nxxq\") pod \"1177173d-34cc-4f8c-9ac9-e2661449305f\" (UID: \"1177173d-34cc-4f8c-9ac9-e2661449305f\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.402735 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1177173d-34cc-4f8c-9ac9-e2661449305f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1177173d-34cc-4f8c-9ac9-e2661449305f" (UID: "1177173d-34cc-4f8c-9ac9-e2661449305f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.403140 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e88366f6-4404-46ae-8538-a7bd3cc9bbb5" (UID: "e88366f6-4404-46ae-8538-a7bd3cc9bbb5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.403637 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1177173d-34cc-4f8c-9ac9-e2661449305f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.403707 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts2nc\" (UniqueName: \"kubernetes.io/projected/173171c8-4339-42ae-a4bc-be3cdc3420be-kube-api-access-ts2nc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.403759 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.403807 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/173171c8-4339-42ae-a4bc-be3cdc3420be-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.405386 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-kube-api-access-xhbpx" (OuterVolumeSpecName: "kube-api-access-xhbpx") pod "e88366f6-4404-46ae-8538-a7bd3cc9bbb5" (UID: "e88366f6-4404-46ae-8538-a7bd3cc9bbb5"). InnerVolumeSpecName "kube-api-access-xhbpx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.405784 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1177173d-34cc-4f8c-9ac9-e2661449305f-kube-api-access-9nxxq" (OuterVolumeSpecName: "kube-api-access-9nxxq") pod "1177173d-34cc-4f8c-9ac9-e2661449305f" (UID: "1177173d-34cc-4f8c-9ac9-e2661449305f"). InnerVolumeSpecName "kube-api-access-9nxxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.504573 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-operator-scripts\") pod \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.504614 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dxmr\" (UniqueName: \"kubernetes.io/projected/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-kube-api-access-7dxmr\") pod \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\" (UID: \"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.504711 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7eb59f9-ff28-4260-834c-dffefdc7db5e-operator-scripts\") pod \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.504766 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxhfp\" (UniqueName: \"kubernetes.io/projected/a7eb59f9-ff28-4260-834c-dffefdc7db5e-kube-api-access-cxhfp\") pod \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\" (UID: \"a7eb59f9-ff28-4260-834c-dffefdc7db5e\") " Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.505102 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhbpx\" (UniqueName: \"kubernetes.io/projected/e88366f6-4404-46ae-8538-a7bd3cc9bbb5-kube-api-access-xhbpx\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.505113 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nxxq\" (UniqueName: \"kubernetes.io/projected/1177173d-34cc-4f8c-9ac9-e2661449305f-kube-api-access-9nxxq\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.506121 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7eb59f9-ff28-4260-834c-dffefdc7db5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a7eb59f9-ff28-4260-834c-dffefdc7db5e" (UID: "a7eb59f9-ff28-4260-834c-dffefdc7db5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.506150 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" (UID: "4dd2a291-3cb5-43ab-8f6c-c7b9517f3455"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.508566 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7eb59f9-ff28-4260-834c-dffefdc7db5e-kube-api-access-cxhfp" (OuterVolumeSpecName: "kube-api-access-cxhfp") pod "a7eb59f9-ff28-4260-834c-dffefdc7db5e" (UID: "a7eb59f9-ff28-4260-834c-dffefdc7db5e"). InnerVolumeSpecName "kube-api-access-cxhfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.509068 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-kube-api-access-7dxmr" (OuterVolumeSpecName: "kube-api-access-7dxmr") pod "4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" (UID: "4dd2a291-3cb5-43ab-8f6c-c7b9517f3455"). InnerVolumeSpecName "kube-api-access-7dxmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.606526 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a7eb59f9-ff28-4260-834c-dffefdc7db5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.606563 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxhfp\" (UniqueName: \"kubernetes.io/projected/a7eb59f9-ff28-4260-834c-dffefdc7db5e-kube-api-access-cxhfp\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.606573 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.606582 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dxmr\" (UniqueName: \"kubernetes.io/projected/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455-kube-api-access-7dxmr\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.762886 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-37f5-account-create-update-m9f9n" event={"ID":"173171c8-4339-42ae-a4bc-be3cdc3420be","Type":"ContainerDied","Data":"7c4080d7c8c1afa345a935b3a4835b168b11a2a295aae0c01fdb0c23529f1f09"} Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.762952 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c4080d7c8c1afa345a935b3a4835b168b11a2a295aae0c01fdb0c23529f1f09" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.763078 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-37f5-account-create-update-m9f9n" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.768089 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-b4thg" event={"ID":"1177173d-34cc-4f8c-9ac9-e2661449305f","Type":"ContainerDied","Data":"7104c25979bd26ec6369c681027cd71cb5bc082d138e613bc650dc1276171d80"} Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.768129 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7104c25979bd26ec6369c681027cd71cb5bc082d138e613bc650dc1276171d80" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.768184 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-b4thg" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.775150 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-znf8s" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.775409 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-znf8s" event={"ID":"e88366f6-4404-46ae-8538-a7bd3cc9bbb5","Type":"ContainerDied","Data":"36085710ec82d4db1a001b4049cf2b01a6b15f0b68599a4a82d5ae3c71100479"} Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.775438 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36085710ec82d4db1a001b4049cf2b01a6b15f0b68599a4a82d5ae3c71100479" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.777801 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-47fe-account-create-update-dpg62" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.777818 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-47fe-account-create-update-dpg62" event={"ID":"4dd2a291-3cb5-43ab-8f6c-c7b9517f3455","Type":"ContainerDied","Data":"38c0f184e5d4b0adbcf0ac12901708db3b7fa08d2231dc3b7a47fc1ee5983b9e"} Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.777882 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38c0f184e5d4b0adbcf0ac12901708db3b7fa08d2231dc3b7a47fc1ee5983b9e" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.785845 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-swzf4" event={"ID":"a7eb59f9-ff28-4260-834c-dffefdc7db5e","Type":"ContainerDied","Data":"b9d78c1c6988255cca42b6a6484f5520c2b0b95ca58c79fb776eb579150357d0"} Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.785876 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9d78c1c6988255cca42b6a6484f5520c2b0b95ca58c79fb776eb579150357d0" Dec 04 12:04:09 crc kubenswrapper[4979]: I1204 12:04:09.785891 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-swzf4" Dec 04 12:04:10 crc kubenswrapper[4979]: I1204 12:04:10.213122 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" path="/var/lib/kubelet/pods/b3c7923a-4a30-4260-ab5a-56d55fbab3e2/volumes" Dec 04 12:04:11 crc kubenswrapper[4979]: I1204 12:04:11.803488 4979 generic.go:334] "Generic (PLEG): container finished" podID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" containerID="3c64b869abf2343d534fe0e3db701fe908e10cca6fa7036f2bb28093bd4e5f7e" exitCode=0 Dec 04 12:04:11 crc kubenswrapper[4979]: I1204 12:04:11.803528 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8kdt5" event={"ID":"f4f85b75-ee47-4c71-822c-e30cd4f3492c","Type":"ContainerDied","Data":"3c64b869abf2343d534fe0e3db701fe908e10cca6fa7036f2bb28093bd4e5f7e"} Dec 04 12:04:12 crc kubenswrapper[4979]: I1204 12:04:12.818515 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4q48m" event={"ID":"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b","Type":"ContainerStarted","Data":"d36d0ca06131ebdb57c910e9f3a30e4390a3daaaeb7dc548462de91ece6eec7c"} Dec 04 12:04:12 crc kubenswrapper[4979]: I1204 12:04:12.844973 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-4q48m" podStartSLOduration=2.823906411 podStartE2EDuration="7.844947263s" podCreationTimestamp="2025-12-04 12:04:05 +0000 UTC" firstStartedPulling="2025-12-04 12:04:06.9755244 +0000 UTC m=+1271.249820204" lastFinishedPulling="2025-12-04 12:04:11.996565212 +0000 UTC m=+1276.270861056" observedRunningTime="2025-12-04 12:04:12.836482873 +0000 UTC m=+1277.110778687" watchObservedRunningTime="2025-12-04 12:04:12.844947263 +0000 UTC m=+1277.119243067" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.197981 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8kdt5" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.364669 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-db-sync-config-data\") pod \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.364869 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h4849\" (UniqueName: \"kubernetes.io/projected/f4f85b75-ee47-4c71-822c-e30cd4f3492c-kube-api-access-h4849\") pod \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.364929 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-combined-ca-bundle\") pod \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.364956 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-config-data\") pod \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\" (UID: \"f4f85b75-ee47-4c71-822c-e30cd4f3492c\") " Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.374630 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4f85b75-ee47-4c71-822c-e30cd4f3492c-kube-api-access-h4849" (OuterVolumeSpecName: "kube-api-access-h4849") pod "f4f85b75-ee47-4c71-822c-e30cd4f3492c" (UID: "f4f85b75-ee47-4c71-822c-e30cd4f3492c"). InnerVolumeSpecName "kube-api-access-h4849". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.386043 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f4f85b75-ee47-4c71-822c-e30cd4f3492c" (UID: "f4f85b75-ee47-4c71-822c-e30cd4f3492c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.398177 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f4f85b75-ee47-4c71-822c-e30cd4f3492c" (UID: "f4f85b75-ee47-4c71-822c-e30cd4f3492c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.409801 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-config-data" (OuterVolumeSpecName: "config-data") pod "f4f85b75-ee47-4c71-822c-e30cd4f3492c" (UID: "f4f85b75-ee47-4c71-822c-e30cd4f3492c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.466762 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h4849\" (UniqueName: \"kubernetes.io/projected/f4f85b75-ee47-4c71-822c-e30cd4f3492c-kube-api-access-h4849\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.466801 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.466812 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.466821 4979 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f4f85b75-ee47-4c71-822c-e30cd4f3492c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.845710 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-8kdt5" event={"ID":"f4f85b75-ee47-4c71-822c-e30cd4f3492c","Type":"ContainerDied","Data":"03d55b25b66ba9dd2d4cfb5fab327d07adc05fda38777b816d5f4be3d200fe8d"} Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.845728 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-8kdt5" Dec 04 12:04:13 crc kubenswrapper[4979]: I1204 12:04:13.845902 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03d55b25b66ba9dd2d4cfb5fab327d07adc05fda38777b816d5f4be3d200fe8d" Dec 04 12:04:13 crc kubenswrapper[4979]: E1204 12:04:13.975611 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4f85b75_ee47_4c71_822c_e30cd4f3492c.slice/crio-03d55b25b66ba9dd2d4cfb5fab327d07adc05fda38777b816d5f4be3d200fe8d\": RecentStats: unable to find data in memory cache]" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.195368 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-9stnk"] Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.195991 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerName="init" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196012 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerName="init" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196023 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e88366f6-4404-46ae-8538-a7bd3cc9bbb5" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196033 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e88366f6-4404-46ae-8538-a7bd3cc9bbb5" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196060 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" containerName="glance-db-sync" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196069 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" containerName="glance-db-sync" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196083 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="173171c8-4339-42ae-a4bc-be3cdc3420be" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196093 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="173171c8-4339-42ae-a4bc-be3cdc3420be" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196106 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerName="dnsmasq-dns" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196115 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerName="dnsmasq-dns" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196137 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196146 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196161 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1177173d-34cc-4f8c-9ac9-e2661449305f" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196170 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1177173d-34cc-4f8c-9ac9-e2661449305f" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196186 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7d638c-e396-403e-974b-864194d22469" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196194 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7d638c-e396-403e-974b-864194d22469" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: E1204 12:04:14.196209 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7eb59f9-ff28-4260-834c-dffefdc7db5e" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196217 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7eb59f9-ff28-4260-834c-dffefdc7db5e" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196423 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e88366f6-4404-46ae-8538-a7bd3cc9bbb5" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196448 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196463 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7d638c-e396-403e-974b-864194d22469" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196473 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c7923a-4a30-4260-ab5a-56d55fbab3e2" containerName="dnsmasq-dns" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196487 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7eb59f9-ff28-4260-834c-dffefdc7db5e" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196499 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" containerName="glance-db-sync" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196531 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1177173d-34cc-4f8c-9ac9-e2661449305f" containerName="mariadb-database-create" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.196543 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="173171c8-4339-42ae-a4bc-be3cdc3420be" containerName="mariadb-account-create-update" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.197621 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.224695 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-9stnk"] Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.281275 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.281362 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.281551 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.281626 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-config\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.281652 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.281676 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5sjz\" (UniqueName: \"kubernetes.io/projected/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-kube-api-access-h5sjz\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.383136 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.383248 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.384006 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.384229 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.384470 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.385192 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.385269 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-config\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.385891 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-config\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.385960 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.385990 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5sjz\" (UniqueName: \"kubernetes.io/projected/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-kube-api-access-h5sjz\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.386666 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.407673 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5sjz\" (UniqueName: \"kubernetes.io/projected/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-kube-api-access-h5sjz\") pod \"dnsmasq-dns-74f6bcbc87-9stnk\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:14 crc kubenswrapper[4979]: I1204 12:04:14.520635 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:15 crc kubenswrapper[4979]: I1204 12:04:15.005529 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-9stnk"] Dec 04 12:04:15 crc kubenswrapper[4979]: I1204 12:04:15.859202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" event={"ID":"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c","Type":"ContainerStarted","Data":"04b28275f2603099b186da06d3d4d6694c0579b01f9ccef38947a72c8fd56340"} Dec 04 12:04:16 crc kubenswrapper[4979]: I1204 12:04:16.867483 4979 generic.go:334] "Generic (PLEG): container finished" podID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerID="6fd6af3b6526b6b8ec951ca793657052f611b231d84998a85705f141fc5d384b" exitCode=0 Dec 04 12:04:16 crc kubenswrapper[4979]: I1204 12:04:16.867575 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" event={"ID":"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c","Type":"ContainerDied","Data":"6fd6af3b6526b6b8ec951ca793657052f611b231d84998a85705f141fc5d384b"} Dec 04 12:04:16 crc kubenswrapper[4979]: I1204 12:04:16.869845 4979 generic.go:334] "Generic (PLEG): container finished" podID="30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" containerID="d36d0ca06131ebdb57c910e9f3a30e4390a3daaaeb7dc548462de91ece6eec7c" exitCode=0 Dec 04 12:04:16 crc kubenswrapper[4979]: I1204 12:04:16.869888 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4q48m" event={"ID":"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b","Type":"ContainerDied","Data":"d36d0ca06131ebdb57c910e9f3a30e4390a3daaaeb7dc548462de91ece6eec7c"} Dec 04 12:04:17 crc kubenswrapper[4979]: I1204 12:04:17.879602 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" event={"ID":"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c","Type":"ContainerStarted","Data":"0d782a1679234ce5db46c213437b695ba7b326a74f0622a424a0921fe6d1b5a9"} Dec 04 12:04:17 crc kubenswrapper[4979]: I1204 12:04:17.905760 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" podStartSLOduration=3.905736585 podStartE2EDuration="3.905736585s" podCreationTimestamp="2025-12-04 12:04:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:17.897965843 +0000 UTC m=+1282.172261667" watchObservedRunningTime="2025-12-04 12:04:17.905736585 +0000 UTC m=+1282.180032399" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.184281 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.254557 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8cqc\" (UniqueName: \"kubernetes.io/projected/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-kube-api-access-f8cqc\") pod \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.254614 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-combined-ca-bundle\") pod \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.254663 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-config-data\") pod \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\" (UID: \"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b\") " Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.262318 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-kube-api-access-f8cqc" (OuterVolumeSpecName: "kube-api-access-f8cqc") pod "30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" (UID: "30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b"). InnerVolumeSpecName "kube-api-access-f8cqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.280576 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" (UID: "30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.321215 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-config-data" (OuterVolumeSpecName: "config-data") pod "30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" (UID: "30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.357147 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8cqc\" (UniqueName: \"kubernetes.io/projected/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-kube-api-access-f8cqc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.357919 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.358295 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.889901 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-4q48m" event={"ID":"30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b","Type":"ContainerDied","Data":"e039cb4eb68ae96961c1acf7d288c4fb5927cecd86df9e89a69b6a829f1e2a3a"} Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.889977 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e039cb4eb68ae96961c1acf7d288c4fb5927cecd86df9e89a69b6a829f1e2a3a" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.889919 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-4q48m" Dec 04 12:04:18 crc kubenswrapper[4979]: I1204 12:04:18.890224 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.197279 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9z5bk"] Dec 04 12:04:19 crc kubenswrapper[4979]: E1204 12:04:19.197742 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" containerName="keystone-db-sync" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.197763 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" containerName="keystone-db-sync" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.197993 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" containerName="keystone-db-sync" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.198974 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.202067 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.202243 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.205703 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-9stnk"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.205862 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zbvzn" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.205870 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.207207 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.218383 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9z5bk"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.234399 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-h5kbj"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.236056 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.267268 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-h5kbj"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.273567 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-fernet-keys\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.273620 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-scripts\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.273656 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-credential-keys\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.273684 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-combined-ca-bundle\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.273734 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs8cb\" (UniqueName: \"kubernetes.io/projected/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-kube-api-access-rs8cb\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.273747 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-config-data\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.366866 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-scv9t"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.368586 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.371759 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.371772 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.371938 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jltrn" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.374851 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.374897 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-fernet-keys\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.374939 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-scripts\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.374959 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-svc\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.374978 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-credential-keys\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375020 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-combined-ca-bundle\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375046 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375069 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-config\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375099 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs8cb\" (UniqueName: \"kubernetes.io/projected/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-kube-api-access-rs8cb\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375115 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375135 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-config-data\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.375197 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9tgp\" (UniqueName: \"kubernetes.io/projected/1b868885-11c9-4c41-a6a5-c5b4560b918a-kube-api-access-q9tgp\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.380386 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-scripts\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.382962 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-credential-keys\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.388468 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-combined-ca-bundle\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.391764 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-fernet-keys\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.399964 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-scv9t"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.407842 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-config-data\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.415607 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs8cb\" (UniqueName: \"kubernetes.io/projected/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-kube-api-access-rs8cb\") pod \"keystone-bootstrap-9z5bk\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.450062 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.458470 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.461361 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.462763 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496746 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9tgp\" (UniqueName: \"kubernetes.io/projected/1b868885-11c9-4c41-a6a5-c5b4560b918a-kube-api-access-q9tgp\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496789 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496840 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-549dz\" (UniqueName: \"kubernetes.io/projected/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-kube-api-access-549dz\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496871 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-svc\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496894 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-config\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496929 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496951 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-config\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.496980 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.497000 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-combined-ca-bundle\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.502092 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.503081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.503739 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.509142 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-svc\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.509241 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-config\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.515358 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.522492 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.558593 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-2465j"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.561205 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.564497 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9tgp\" (UniqueName: \"kubernetes.io/projected/1b868885-11c9-4c41-a6a5-c5b4560b918a-kube-api-access-q9tgp\") pod \"dnsmasq-dns-847c4cc679-h5kbj\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.564910 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.565188 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zf9zs" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.565253 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.573447 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.582497 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xmtnh"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.584660 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.587620 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zgx4w" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.587957 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.595211 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2465j"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599560 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-scripts\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599607 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxbl5\" (UniqueName: \"kubernetes.io/projected/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-kube-api-access-bxbl5\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599644 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-combined-ca-bundle\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599683 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599704 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-config-data\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599727 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-run-httpd\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599753 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599781 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-549dz\" (UniqueName: \"kubernetes.io/projected/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-kube-api-access-549dz\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599805 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-config\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.599848 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-log-httpd\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.611233 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xmtnh"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.613992 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-config\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.617451 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-combined-ca-bundle\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.621553 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-549dz\" (UniqueName: \"kubernetes.io/projected/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-kube-api-access-549dz\") pod \"neutron-db-sync-scv9t\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.660291 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-h5kbj"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.731534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-combined-ca-bundle\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.731808 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-95q7j"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.731870 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.731908 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-config-data\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.731947 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-run-httpd\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.731982 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp8l4\" (UniqueName: \"kubernetes.io/projected/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-kube-api-access-qp8l4\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732009 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-config-data\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732063 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-etc-machine-id\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732082 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-db-sync-config-data\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732101 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-db-sync-config-data\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732134 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckxtc\" (UniqueName: \"kubernetes.io/projected/e9ceee6e-4685-4dc1-8222-57cc6d51a337-kube-api-access-ckxtc\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732164 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-scripts\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732203 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-log-httpd\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732229 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-combined-ca-bundle\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732250 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-scripts\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.732282 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxbl5\" (UniqueName: \"kubernetes.io/projected/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-kube-api-access-bxbl5\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.738499 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-95q7j"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.738913 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.741401 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.741757 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-run-httpd\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.749877 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-log-httpd\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.754837 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.763882 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-config-data\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.767291 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-s2g8b"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.772031 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.773707 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.776773 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bbh4d" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.777132 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.777262 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.784086 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-scripts\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.786393 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxbl5\" (UniqueName: \"kubernetes.io/projected/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-kube-api-access-bxbl5\") pod \"ceilometer-0\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.790284 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.810230 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s2g8b"] Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839224 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-config\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839289 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839355 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp8l4\" (UniqueName: \"kubernetes.io/projected/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-kube-api-access-qp8l4\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839387 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-config-data\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839412 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-etc-machine-id\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839738 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-db-sync-config-data\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839761 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-db-sync-config-data\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839791 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckxtc\" (UniqueName: \"kubernetes.io/projected/e9ceee6e-4685-4dc1-8222-57cc6d51a337-kube-api-access-ckxtc\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839821 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-scripts\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839870 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-combined-ca-bundle\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839889 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839939 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wzg4\" (UniqueName: \"kubernetes.io/projected/edba8570-7f7d-4c8e-8917-862f5dcc59ba-kube-api-access-9wzg4\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839955 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.839974 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.840010 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-combined-ca-bundle\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.844048 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-etc-machine-id\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.844957 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-scripts\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.845003 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-db-sync-config-data\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.845098 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-combined-ca-bundle\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.846982 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-combined-ca-bundle\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.848460 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-db-sync-config-data\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.848567 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-config-data\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.874248 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp8l4\" (UniqueName: \"kubernetes.io/projected/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-kube-api-access-qp8l4\") pod \"cinder-db-sync-2465j\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.890799 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckxtc\" (UniqueName: \"kubernetes.io/projected/e9ceee6e-4685-4dc1-8222-57cc6d51a337-kube-api-access-ckxtc\") pod \"barbican-db-sync-xmtnh\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943138 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943192 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-config-data\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943213 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-logs\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943277 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-scripts\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943338 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943376 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943391 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wzg4\" (UniqueName: \"kubernetes.io/projected/edba8570-7f7d-4c8e-8917-862f5dcc59ba-kube-api-access-9wzg4\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943411 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943454 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-config\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943472 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-combined-ca-bundle\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.943497 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfrbc\" (UniqueName: \"kubernetes.io/projected/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-kube-api-access-hfrbc\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.944342 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.945444 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.961072 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.966625 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.966691 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-config\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:19 crc kubenswrapper[4979]: I1204 12:04:19.992081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wzg4\" (UniqueName: \"kubernetes.io/projected/edba8570-7f7d-4c8e-8917-862f5dcc59ba-kube-api-access-9wzg4\") pod \"dnsmasq-dns-785d8bcb8c-95q7j\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.045759 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-combined-ca-bundle\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.045824 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfrbc\" (UniqueName: \"kubernetes.io/projected/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-kube-api-access-hfrbc\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.045862 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-config-data\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.045880 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-logs\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.045980 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-scripts\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.049406 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-logs\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.061330 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-scripts\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.064058 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-config-data\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.064081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-combined-ca-bundle\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.077438 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2465j" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.092749 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.093721 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfrbc\" (UniqueName: \"kubernetes.io/projected/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-kube-api-access-hfrbc\") pod \"placement-db-sync-s2g8b\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.126834 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.137710 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.142766 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9z5bk"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.230669 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-h5kbj"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.315286 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.320274 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.323966 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-ksp64" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.324854 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.325913 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.326132 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.337271 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.388614 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.390159 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.392938 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.397041 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.405412 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459554 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459597 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-scripts\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459643 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qshf\" (UniqueName: \"kubernetes.io/projected/008a3fb2-b015-40d2-9836-79e57113efc2-kube-api-access-4qshf\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459661 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459721 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459742 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-config-data\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459771 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.459798 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-logs\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.461647 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-scv9t"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.561711 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562042 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qshf\" (UniqueName: \"kubernetes.io/projected/008a3fb2-b015-40d2-9836-79e57113efc2-kube-api-access-4qshf\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562067 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rptx\" (UniqueName: \"kubernetes.io/projected/014a18ee-7808-4217-b04c-2857c34d6b0d-kube-api-access-9rptx\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562085 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562140 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562164 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562188 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562210 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-config-data\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562229 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562249 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562265 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-logs\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562283 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-logs\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562330 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562386 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.562656 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.563181 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-logs\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.566875 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.567247 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-scripts\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.567271 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.567641 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.570947 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-config-data\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.574102 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.578137 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-scripts\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.597621 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qshf\" (UniqueName: \"kubernetes.io/projected/008a3fb2-b015-40d2-9836-79e57113efc2-kube-api-access-4qshf\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.608073 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.627608 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.648583 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.669145 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.669902 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.669936 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-logs\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.670004 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.670030 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.670118 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.670168 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rptx\" (UniqueName: \"kubernetes.io/projected/014a18ee-7808-4217-b04c-2857c34d6b0d-kube-api-access-9rptx\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.670248 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.670829 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.674067 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-scripts\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.675102 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.676103 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-config-data\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.676838 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-logs\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.677985 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.678570 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.690632 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rptx\" (UniqueName: \"kubernetes.io/projected/014a18ee-7808-4217-b04c-2857c34d6b0d-kube-api-access-9rptx\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.719579 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.769324 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.774529 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xmtnh"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.807044 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-2465j"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.812445 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-95q7j"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.900452 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-s2g8b"] Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.935353 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s2g8b" event={"ID":"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112","Type":"ContainerStarted","Data":"d3cc2b03b2165a86339a57d1724f4f0b1b40c9fe4538b5262557bd24b0d85e71"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.940418 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2465j" event={"ID":"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6","Type":"ContainerStarted","Data":"5c46d7ea59c4d94ad039c40d3eec69ed33119159d134390ee76c47be56200ea7"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.943714 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-scv9t" event={"ID":"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9","Type":"ContainerStarted","Data":"aed86b6e4e107649f8fb95f66dcae276fcc47522a30318a9927a27f1a5c5a593"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.943739 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-scv9t" event={"ID":"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9","Type":"ContainerStarted","Data":"35478a4648d79d8ec051ef0badd4a1cca8a17a5324912142ebd7152609afd2f1"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.946283 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" event={"ID":"1b868885-11c9-4c41-a6a5-c5b4560b918a","Type":"ContainerStarted","Data":"0f719c0a58c82f2787705154c4c2cd9a6e7a1415c962828975086f620fb69e8e"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.947719 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9z5bk" event={"ID":"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9","Type":"ContainerStarted","Data":"bee5a7aae3b63933dd5a1e0a68fdee80d8892b52496944e2423977d5e0b997af"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.947778 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9z5bk" event={"ID":"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9","Type":"ContainerStarted","Data":"6709f9e7ab14ffc9c0bfea7643eeb40f8eed105d73354548b3f8c76c05b1108e"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.949004 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xmtnh" event={"ID":"e9ceee6e-4685-4dc1-8222-57cc6d51a337","Type":"ContainerStarted","Data":"41686986fde3ea66945072a0b937d7b775a014661a9f06b74fa339a072a4c98d"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.950101 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerStarted","Data":"fa83fc2b3c315f66cb7005de5f65fd09170a5b16c1264ccd42fa5c54bf025c0d"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.951993 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" event={"ID":"edba8570-7f7d-4c8e-8917-862f5dcc59ba","Type":"ContainerStarted","Data":"0432b663e4218badb689c2c4381abbe2e497cb7c44008ca8322c11ab2a547f26"} Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.952226 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerName="dnsmasq-dns" containerID="cri-o://0d782a1679234ce5db46c213437b695ba7b326a74f0622a424a0921fe6d1b5a9" gracePeriod=10 Dec 04 12:04:20 crc kubenswrapper[4979]: I1204 12:04:20.964078 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-scv9t" podStartSLOduration=1.964058039 podStartE2EDuration="1.964058039s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:20.957365267 +0000 UTC m=+1285.231661081" watchObservedRunningTime="2025-12-04 12:04:20.964058039 +0000 UTC m=+1285.238353853" Dec 04 12:04:21 crc kubenswrapper[4979]: I1204 12:04:21.294597 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9z5bk" podStartSLOduration=2.29458048 podStartE2EDuration="2.29458048s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:20.982801748 +0000 UTC m=+1285.257097552" watchObservedRunningTime="2025-12-04 12:04:21.29458048 +0000 UTC m=+1285.568876284" Dec 04 12:04:21 crc kubenswrapper[4979]: I1204 12:04:21.297729 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:21 crc kubenswrapper[4979]: I1204 12:04:21.455731 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.058169 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.082793 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.084279 4979 generic.go:334] "Generic (PLEG): container finished" podID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerID="8145e99f3ce77a36bd15c0b98448ad4bcb34344db4310d820df92958b5118f21" exitCode=0 Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.084396 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" event={"ID":"edba8570-7f7d-4c8e-8917-862f5dcc59ba","Type":"ContainerDied","Data":"8145e99f3ce77a36bd15c0b98448ad4bcb34344db4310d820df92958b5118f21"} Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.092131 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"014a18ee-7808-4217-b04c-2857c34d6b0d","Type":"ContainerStarted","Data":"1870625d52aec2fbf5770b4ac901d061012ed9f9da8e4109bda5c110f6c31b5f"} Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.115491 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.138497 4979 generic.go:334] "Generic (PLEG): container finished" podID="1b868885-11c9-4c41-a6a5-c5b4560b918a" containerID="da1d036972eae31561fa4d11e55bb90e4acae8318553a69182088d8a9f3ebb21" exitCode=0 Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.138582 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" event={"ID":"1b868885-11c9-4c41-a6a5-c5b4560b918a","Type":"ContainerDied","Data":"da1d036972eae31561fa4d11e55bb90e4acae8318553a69182088d8a9f3ebb21"} Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.143901 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.159828 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"008a3fb2-b015-40d2-9836-79e57113efc2","Type":"ContainerStarted","Data":"7a8f48baf5b78b0bfde8a8eff5f6b26f4e28534451e400dff3819d336e1fc74f"} Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.218963 4979 generic.go:334] "Generic (PLEG): container finished" podID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerID="0d782a1679234ce5db46c213437b695ba7b326a74f0622a424a0921fe6d1b5a9" exitCode=0 Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.219520 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.258420 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-9stnk" event={"ID":"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c","Type":"ContainerDied","Data":"0d782a1679234ce5db46c213437b695ba7b326a74f0622a424a0921fe6d1b5a9"} Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.258461 4979 scope.go:117] "RemoveContainer" containerID="0d782a1679234ce5db46c213437b695ba7b326a74f0622a424a0921fe6d1b5a9" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.259111 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-swift-storage-0\") pod \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.259163 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-config\") pod \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.259203 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-nb\") pod \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.259228 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5sjz\" (UniqueName: \"kubernetes.io/projected/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-kube-api-access-h5sjz\") pod \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.272333 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-svc\") pod \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.272420 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-sb\") pod \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\" (UID: \"838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.279342 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-kube-api-access-h5sjz" (OuterVolumeSpecName: "kube-api-access-h5sjz") pod "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" (UID: "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c"). InnerVolumeSpecName "kube-api-access-h5sjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.371200 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" (UID: "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.372953 4979 scope.go:117] "RemoveContainer" containerID="6fd6af3b6526b6b8ec951ca793657052f611b231d84998a85705f141fc5d384b" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.374995 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5sjz\" (UniqueName: \"kubernetes.io/projected/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-kube-api-access-h5sjz\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.375080 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.430292 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" (UID: "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.467061 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" (UID: "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.476589 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.476612 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.480748 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-config" (OuterVolumeSpecName: "config") pod "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" (UID: "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.490502 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" (UID: "838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.578353 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.578427 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.627669 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.646095 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-9stnk"] Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.663362 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-9stnk"] Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.679264 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-nb\") pod \"1b868885-11c9-4c41-a6a5-c5b4560b918a\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.679629 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-sb\") pod \"1b868885-11c9-4c41-a6a5-c5b4560b918a\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.679669 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-swift-storage-0\") pod \"1b868885-11c9-4c41-a6a5-c5b4560b918a\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.679768 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-config\") pod \"1b868885-11c9-4c41-a6a5-c5b4560b918a\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.679961 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9tgp\" (UniqueName: \"kubernetes.io/projected/1b868885-11c9-4c41-a6a5-c5b4560b918a-kube-api-access-q9tgp\") pod \"1b868885-11c9-4c41-a6a5-c5b4560b918a\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.680024 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-svc\") pod \"1b868885-11c9-4c41-a6a5-c5b4560b918a\" (UID: \"1b868885-11c9-4c41-a6a5-c5b4560b918a\") " Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.697581 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b868885-11c9-4c41-a6a5-c5b4560b918a-kube-api-access-q9tgp" (OuterVolumeSpecName: "kube-api-access-q9tgp") pod "1b868885-11c9-4c41-a6a5-c5b4560b918a" (UID: "1b868885-11c9-4c41-a6a5-c5b4560b918a"). InnerVolumeSpecName "kube-api-access-q9tgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.712990 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1b868885-11c9-4c41-a6a5-c5b4560b918a" (UID: "1b868885-11c9-4c41-a6a5-c5b4560b918a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.716898 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-config" (OuterVolumeSpecName: "config") pod "1b868885-11c9-4c41-a6a5-c5b4560b918a" (UID: "1b868885-11c9-4c41-a6a5-c5b4560b918a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.739048 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1b868885-11c9-4c41-a6a5-c5b4560b918a" (UID: "1b868885-11c9-4c41-a6a5-c5b4560b918a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.745779 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b868885-11c9-4c41-a6a5-c5b4560b918a" (UID: "1b868885-11c9-4c41-a6a5-c5b4560b918a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.754101 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1b868885-11c9-4c41-a6a5-c5b4560b918a" (UID: "1b868885-11c9-4c41-a6a5-c5b4560b918a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.785435 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.785465 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.785476 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.785487 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.785496 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9tgp\" (UniqueName: \"kubernetes.io/projected/1b868885-11c9-4c41-a6a5-c5b4560b918a-kube-api-access-q9tgp\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:22 crc kubenswrapper[4979]: I1204 12:04:22.785504 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b868885-11c9-4c41-a6a5-c5b4560b918a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.255371 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.255411 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-h5kbj" event={"ID":"1b868885-11c9-4c41-a6a5-c5b4560b918a","Type":"ContainerDied","Data":"0f719c0a58c82f2787705154c4c2cd9a6e7a1415c962828975086f620fb69e8e"} Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.255481 4979 scope.go:117] "RemoveContainer" containerID="da1d036972eae31561fa4d11e55bb90e4acae8318553a69182088d8a9f3ebb21" Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.259396 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"008a3fb2-b015-40d2-9836-79e57113efc2","Type":"ContainerStarted","Data":"ccf2c599204da74e3bee6b9dfecaa29381899e2aed874f6fbcef0e9e54dc9839"} Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.263636 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" event={"ID":"edba8570-7f7d-4c8e-8917-862f5dcc59ba","Type":"ContainerStarted","Data":"b47af92c762c0b6617f5114d10bd61c6e56f191d7fa672d3c0db1f97f616a2b8"} Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.263727 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.282879 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"014a18ee-7808-4217-b04c-2857c34d6b0d","Type":"ContainerStarted","Data":"192375eee4b5d01e4694d0ec9847562d172334163d40d6c0f9ff8761ba1fc7a6"} Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.288189 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" podStartSLOduration=4.288168466 podStartE2EDuration="4.288168466s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:23.28279406 +0000 UTC m=+1287.557089864" watchObservedRunningTime="2025-12-04 12:04:23.288168466 +0000 UTC m=+1287.562464270" Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.383631 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-h5kbj"] Dec 04 12:04:23 crc kubenswrapper[4979]: I1204 12:04:23.395706 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-h5kbj"] Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.208640 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b868885-11c9-4c41-a6a5-c5b4560b918a" path="/var/lib/kubelet/pods/1b868885-11c9-4c41-a6a5-c5b4560b918a/volumes" Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.209401 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" path="/var/lib/kubelet/pods/838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c/volumes" Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.305952 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"014a18ee-7808-4217-b04c-2857c34d6b0d","Type":"ContainerStarted","Data":"3942f5955323529add170680571dc4cf8b61dcfc9c9e2af9f7a0ac61b7db5632"} Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.306112 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-log" containerID="cri-o://192375eee4b5d01e4694d0ec9847562d172334163d40d6c0f9ff8761ba1fc7a6" gracePeriod=30 Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.306280 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-httpd" containerID="cri-o://3942f5955323529add170680571dc4cf8b61dcfc9c9e2af9f7a0ac61b7db5632" gracePeriod=30 Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.319756 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-log" containerID="cri-o://ccf2c599204da74e3bee6b9dfecaa29381899e2aed874f6fbcef0e9e54dc9839" gracePeriod=30 Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.319860 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-httpd" containerID="cri-o://cf9068416f8b059541200f4a2196123cb243bb95381d03f2a7cad3623c7e3f72" gracePeriod=30 Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.319767 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"008a3fb2-b015-40d2-9836-79e57113efc2","Type":"ContainerStarted","Data":"cf9068416f8b059541200f4a2196123cb243bb95381d03f2a7cad3623c7e3f72"} Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.330770 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.330749803 podStartE2EDuration="5.330749803s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:24.32804635 +0000 UTC m=+1288.602342154" watchObservedRunningTime="2025-12-04 12:04:24.330749803 +0000 UTC m=+1288.605045607" Dec 04 12:04:24 crc kubenswrapper[4979]: I1204 12:04:24.364740 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.364716826 podStartE2EDuration="5.364716826s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:24.355637408 +0000 UTC m=+1288.629933222" watchObservedRunningTime="2025-12-04 12:04:24.364716826 +0000 UTC m=+1288.639012630" Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.327688 4979 generic.go:334] "Generic (PLEG): container finished" podID="008a3fb2-b015-40d2-9836-79e57113efc2" containerID="cf9068416f8b059541200f4a2196123cb243bb95381d03f2a7cad3623c7e3f72" exitCode=0 Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.328028 4979 generic.go:334] "Generic (PLEG): container finished" podID="008a3fb2-b015-40d2-9836-79e57113efc2" containerID="ccf2c599204da74e3bee6b9dfecaa29381899e2aed874f6fbcef0e9e54dc9839" exitCode=143 Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.327780 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"008a3fb2-b015-40d2-9836-79e57113efc2","Type":"ContainerDied","Data":"cf9068416f8b059541200f4a2196123cb243bb95381d03f2a7cad3623c7e3f72"} Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.328072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"008a3fb2-b015-40d2-9836-79e57113efc2","Type":"ContainerDied","Data":"ccf2c599204da74e3bee6b9dfecaa29381899e2aed874f6fbcef0e9e54dc9839"} Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.329717 4979 generic.go:334] "Generic (PLEG): container finished" podID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerID="3942f5955323529add170680571dc4cf8b61dcfc9c9e2af9f7a0ac61b7db5632" exitCode=0 Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.329765 4979 generic.go:334] "Generic (PLEG): container finished" podID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerID="192375eee4b5d01e4694d0ec9847562d172334163d40d6c0f9ff8761ba1fc7a6" exitCode=143 Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.329775 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"014a18ee-7808-4217-b04c-2857c34d6b0d","Type":"ContainerDied","Data":"3942f5955323529add170680571dc4cf8b61dcfc9c9e2af9f7a0ac61b7db5632"} Dec 04 12:04:25 crc kubenswrapper[4979]: I1204 12:04:25.329811 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"014a18ee-7808-4217-b04c-2857c34d6b0d","Type":"ContainerDied","Data":"192375eee4b5d01e4694d0ec9847562d172334163d40d6c0f9ff8761ba1fc7a6"} Dec 04 12:04:26 crc kubenswrapper[4979]: I1204 12:04:26.357002 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" containerID="bee5a7aae3b63933dd5a1e0a68fdee80d8892b52496944e2423977d5e0b997af" exitCode=0 Dec 04 12:04:26 crc kubenswrapper[4979]: I1204 12:04:26.357362 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9z5bk" event={"ID":"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9","Type":"ContainerDied","Data":"bee5a7aae3b63933dd5a1e0a68fdee80d8892b52496944e2423977d5e0b997af"} Dec 04 12:04:28 crc kubenswrapper[4979]: I1204 12:04:28.041015 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:04:28 crc kubenswrapper[4979]: I1204 12:04:28.041082 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:04:30 crc kubenswrapper[4979]: I1204 12:04:30.129571 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:04:30 crc kubenswrapper[4979]: I1204 12:04:30.210285 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rmf45"] Dec 04 12:04:30 crc kubenswrapper[4979]: I1204 12:04:30.210514 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="dnsmasq-dns" containerID="cri-o://c8be71aa2ab63e7bba5003b34e684ff345e36633a9b0398fbf6538afd399a7a6" gracePeriod=10 Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.061495 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.067143 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146240 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-config-data\") pod \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146311 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-logs\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146355 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-scripts\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146389 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146461 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rptx\" (UniqueName: \"kubernetes.io/projected/014a18ee-7808-4217-b04c-2857c34d6b0d-kube-api-access-9rptx\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146508 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-fernet-keys\") pod \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146556 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs8cb\" (UniqueName: \"kubernetes.io/projected/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-kube-api-access-rs8cb\") pod \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146582 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-combined-ca-bundle\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146621 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-combined-ca-bundle\") pod \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146644 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-httpd-run\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146670 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-credential-keys\") pod \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146693 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-internal-tls-certs\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146719 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-scripts\") pod \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\" (UID: \"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.146740 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-config-data\") pod \"014a18ee-7808-4217-b04c-2857c34d6b0d\" (UID: \"014a18ee-7808-4217-b04c-2857c34d6b0d\") " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.148119 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.153050 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-logs" (OuterVolumeSpecName: "logs") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.153974 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/014a18ee-7808-4217-b04c-2857c34d6b0d-kube-api-access-9rptx" (OuterVolumeSpecName: "kube-api-access-9rptx") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "kube-api-access-9rptx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.154120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-scripts" (OuterVolumeSpecName: "scripts") pod "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" (UID: "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.162954 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" (UID: "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.163907 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-kube-api-access-rs8cb" (OuterVolumeSpecName: "kube-api-access-rs8cb") pod "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" (UID: "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9"). InnerVolumeSpecName "kube-api-access-rs8cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.171539 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-scripts" (OuterVolumeSpecName: "scripts") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.181934 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" (UID: "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.190205 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-config-data" (OuterVolumeSpecName: "config-data") pod "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" (UID: "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.194593 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.197738 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" (UID: "bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.207600 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252433 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252672 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252683 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252692 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252708 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252719 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rptx\" (UniqueName: \"kubernetes.io/projected/014a18ee-7808-4217-b04c-2857c34d6b0d-kube-api-access-9rptx\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252730 4979 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252740 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs8cb\" (UniqueName: \"kubernetes.io/projected/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-kube-api-access-rs8cb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252748 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252756 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252765 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/014a18ee-7808-4217-b04c-2857c34d6b0d-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.252773 4979 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.256702 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-config-data" (OuterVolumeSpecName: "config-data") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.272899 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.294683 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "014a18ee-7808-4217-b04c-2857c34d6b0d" (UID: "014a18ee-7808-4217-b04c-2857c34d6b0d"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.354076 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.354203 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/014a18ee-7808-4217-b04c-2857c34d6b0d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.354221 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.408792 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"014a18ee-7808-4217-b04c-2857c34d6b0d","Type":"ContainerDied","Data":"1870625d52aec2fbf5770b4ac901d061012ed9f9da8e4109bda5c110f6c31b5f"} Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.408822 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.408839 4979 scope.go:117] "RemoveContainer" containerID="3942f5955323529add170680571dc4cf8b61dcfc9c9e2af9f7a0ac61b7db5632" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.413669 4979 generic.go:334] "Generic (PLEG): container finished" podID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerID="c8be71aa2ab63e7bba5003b34e684ff345e36633a9b0398fbf6538afd399a7a6" exitCode=0 Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.413745 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" event={"ID":"333d2328-de05-4d66-a8b1-d0585ddf5f32","Type":"ContainerDied","Data":"c8be71aa2ab63e7bba5003b34e684ff345e36633a9b0398fbf6538afd399a7a6"} Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.415315 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9z5bk" event={"ID":"bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9","Type":"ContainerDied","Data":"6709f9e7ab14ffc9c0bfea7643eeb40f8eed105d73354548b3f8c76c05b1108e"} Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.415346 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6709f9e7ab14ffc9c0bfea7643eeb40f8eed105d73354548b3f8c76c05b1108e" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.415370 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9z5bk" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.445280 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.460139 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475237 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:31 crc kubenswrapper[4979]: E1204 12:04:31.475683 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" containerName="keystone-bootstrap" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475701 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" containerName="keystone-bootstrap" Dec 04 12:04:31 crc kubenswrapper[4979]: E1204 12:04:31.475719 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-log" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475727 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-log" Dec 04 12:04:31 crc kubenswrapper[4979]: E1204 12:04:31.475742 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b868885-11c9-4c41-a6a5-c5b4560b918a" containerName="init" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475749 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b868885-11c9-4c41-a6a5-c5b4560b918a" containerName="init" Dec 04 12:04:31 crc kubenswrapper[4979]: E1204 12:04:31.475774 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerName="init" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475782 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerName="init" Dec 04 12:04:31 crc kubenswrapper[4979]: E1204 12:04:31.475799 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerName="dnsmasq-dns" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475807 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerName="dnsmasq-dns" Dec 04 12:04:31 crc kubenswrapper[4979]: E1204 12:04:31.475820 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-httpd" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.475828 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-httpd" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.476040 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-httpd" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.476063 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" containerName="glance-log" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.476077 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="838f97d0-bda8-4d34-a8a7-0efb9d4c3c9c" containerName="dnsmasq-dns" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.476086 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" containerName="keystone-bootstrap" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.476096 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b868885-11c9-4c41-a6a5-c5b4560b918a" containerName="init" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.476994 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.482513 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.485187 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.487580 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660416 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-logs\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660485 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660526 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660570 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660647 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skg7k\" (UniqueName: \"kubernetes.io/projected/95a0e5e3-658f-4d68-9387-34966c2a8e1c-kube-api-access-skg7k\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660672 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660699 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.660767 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.762611 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.762951 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-logs\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.762989 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763068 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763133 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skg7k\" (UniqueName: \"kubernetes.io/projected/95a0e5e3-658f-4d68-9387-34966c2a8e1c-kube-api-access-skg7k\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763158 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763186 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763577 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-logs\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763841 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.763879 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.767787 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.768750 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.769325 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.777745 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.784428 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skg7k\" (UniqueName: \"kubernetes.io/projected/95a0e5e3-658f-4d68-9387-34966c2a8e1c-kube-api-access-skg7k\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:31 crc kubenswrapper[4979]: I1204 12:04:31.799821 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.102196 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.211546 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="014a18ee-7808-4217-b04c-2857c34d6b0d" path="/var/lib/kubelet/pods/014a18ee-7808-4217-b04c-2857c34d6b0d/volumes" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.223100 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9z5bk"] Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.230347 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9z5bk"] Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.331469 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-hfg44"] Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.332679 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.337203 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.337203 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.337277 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.337341 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.337528 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zbvzn" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.342684 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hfg44"] Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.482532 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-config-data\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.482578 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-combined-ca-bundle\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.486771 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-credential-keys\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.486911 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-fernet-keys\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.486990 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvssz\" (UniqueName: \"kubernetes.io/projected/8e3201e7-fdd0-49c2-93bc-96211c54e55a-kube-api-access-zvssz\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.487017 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-scripts\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.589128 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-config-data\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.589187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-combined-ca-bundle\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.589254 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-credential-keys\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.589342 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-fernet-keys\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.589401 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvssz\" (UniqueName: \"kubernetes.io/projected/8e3201e7-fdd0-49c2-93bc-96211c54e55a-kube-api-access-zvssz\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.589423 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-scripts\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.594558 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-fernet-keys\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.596958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-combined-ca-bundle\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.597637 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-config-data\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.598026 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-credential-keys\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.599167 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-scripts\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.612281 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvssz\" (UniqueName: \"kubernetes.io/projected/8e3201e7-fdd0-49c2-93bc-96211c54e55a-kube-api-access-zvssz\") pod \"keystone-bootstrap-hfg44\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:32 crc kubenswrapper[4979]: I1204 12:04:32.659171 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:34 crc kubenswrapper[4979]: I1204 12:04:34.212463 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9" path="/var/lib/kubelet/pods/bc29ab35-d21d-4f83-b1b7-94fd74bf7fb9/volumes" Dec 04 12:04:38 crc kubenswrapper[4979]: I1204 12:04:38.188880 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.497133 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.502839 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.502962 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"008a3fb2-b015-40d2-9836-79e57113efc2","Type":"ContainerDied","Data":"7a8f48baf5b78b0bfde8a8eff5f6b26f4e28534451e400dff3819d336e1fc74f"} Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.507011 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.507525 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" event={"ID":"333d2328-de05-4d66-a8b1-d0585ddf5f32","Type":"ContainerDied","Data":"a6d55e841376361a09ee4675cf285d5ee18832620f9e4e42f8359315c118f705"} Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.644773 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-svc\") pod \"333d2328-de05-4d66-a8b1-d0585ddf5f32\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.644863 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-httpd-run\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.644896 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-config\") pod \"333d2328-de05-4d66-a8b1-d0585ddf5f32\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.644951 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-combined-ca-bundle\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.644996 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-scripts\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645062 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-swift-storage-0\") pod \"333d2328-de05-4d66-a8b1-d0585ddf5f32\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645238 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmrq2\" (UniqueName: \"kubernetes.io/projected/333d2328-de05-4d66-a8b1-d0585ddf5f32-kube-api-access-dmrq2\") pod \"333d2328-de05-4d66-a8b1-d0585ddf5f32\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645287 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qshf\" (UniqueName: \"kubernetes.io/projected/008a3fb2-b015-40d2-9836-79e57113efc2-kube-api-access-4qshf\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645334 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645400 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-public-tls-certs\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645444 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-logs\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645456 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645466 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-nb\") pod \"333d2328-de05-4d66-a8b1-d0585ddf5f32\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645540 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-sb\") pod \"333d2328-de05-4d66-a8b1-d0585ddf5f32\" (UID: \"333d2328-de05-4d66-a8b1-d0585ddf5f32\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.645565 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-config-data\") pod \"008a3fb2-b015-40d2-9836-79e57113efc2\" (UID: \"008a3fb2-b015-40d2-9836-79e57113efc2\") " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.646132 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.648396 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-logs" (OuterVolumeSpecName: "logs") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.655518 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.657033 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/333d2328-de05-4d66-a8b1-d0585ddf5f32-kube-api-access-dmrq2" (OuterVolumeSpecName: "kube-api-access-dmrq2") pod "333d2328-de05-4d66-a8b1-d0585ddf5f32" (UID: "333d2328-de05-4d66-a8b1-d0585ddf5f32"). InnerVolumeSpecName "kube-api-access-dmrq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.658030 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-scripts" (OuterVolumeSpecName: "scripts") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.660994 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/008a3fb2-b015-40d2-9836-79e57113efc2-kube-api-access-4qshf" (OuterVolumeSpecName: "kube-api-access-4qshf") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "kube-api-access-4qshf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.682690 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.699012 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "333d2328-de05-4d66-a8b1-d0585ddf5f32" (UID: "333d2328-de05-4d66-a8b1-d0585ddf5f32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.705868 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "333d2328-de05-4d66-a8b1-d0585ddf5f32" (UID: "333d2328-de05-4d66-a8b1-d0585ddf5f32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.706222 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "333d2328-de05-4d66-a8b1-d0585ddf5f32" (UID: "333d2328-de05-4d66-a8b1-d0585ddf5f32"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.711041 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-config-data" (OuterVolumeSpecName: "config-data") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.713438 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "008a3fb2-b015-40d2-9836-79e57113efc2" (UID: "008a3fb2-b015-40d2-9836-79e57113efc2"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.718282 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "333d2328-de05-4d66-a8b1-d0585ddf5f32" (UID: "333d2328-de05-4d66-a8b1-d0585ddf5f32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.727130 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-config" (OuterVolumeSpecName: "config") pod "333d2328-de05-4d66-a8b1-d0585ddf5f32" (UID: "333d2328-de05-4d66-a8b1-d0585ddf5f32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747603 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/008a3fb2-b015-40d2-9836-79e57113efc2-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747637 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747648 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747657 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747667 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747674 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747682 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747690 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747698 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/333d2328-de05-4d66-a8b1-d0585ddf5f32-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747707 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmrq2\" (UniqueName: \"kubernetes.io/projected/333d2328-de05-4d66-a8b1-d0585ddf5f32-kube-api-access-dmrq2\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747714 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qshf\" (UniqueName: \"kubernetes.io/projected/008a3fb2-b015-40d2-9836-79e57113efc2-kube-api-access-4qshf\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747772 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.747782 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/008a3fb2-b015-40d2-9836-79e57113efc2-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.776729 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.842351 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.849223 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.858979 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.867888 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:39 crc kubenswrapper[4979]: E1204 12:04:39.868268 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-log" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868283 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-log" Dec 04 12:04:39 crc kubenswrapper[4979]: E1204 12:04:39.868292 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="dnsmasq-dns" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868311 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="dnsmasq-dns" Dec 04 12:04:39 crc kubenswrapper[4979]: E1204 12:04:39.868334 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-httpd" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868340 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-httpd" Dec 04 12:04:39 crc kubenswrapper[4979]: E1204 12:04:39.868351 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="init" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868357 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="init" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868546 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="dnsmasq-dns" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868561 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-httpd" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.868575 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" containerName="glance-log" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.872307 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.875316 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.887218 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 12:04:39 crc kubenswrapper[4979]: I1204 12:04:39.887574 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052212 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-scripts\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052594 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-config-data\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052625 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlf5t\" (UniqueName: \"kubernetes.io/projected/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-kube-api-access-qlf5t\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052650 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052678 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052712 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052757 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.052785 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-logs\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154471 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-logs\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154581 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-scripts\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154658 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-config-data\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154692 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlf5t\" (UniqueName: \"kubernetes.io/projected/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-kube-api-access-qlf5t\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154724 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154758 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.154798 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.155243 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.156363 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-logs\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.156475 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.159277 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.160925 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-scripts\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.161458 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-config-data\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.172651 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.178156 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlf5t\" (UniqueName: \"kubernetes.io/projected/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-kube-api-access-qlf5t\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.193658 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.213997 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="008a3fb2-b015-40d2-9836-79e57113efc2" path="/var/lib/kubelet/pods/008a3fb2-b015-40d2-9836-79e57113efc2/volumes" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.494470 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.514907 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.539588 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rmf45"] Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.546215 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rmf45"] Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.659102 4979 scope.go:117] "RemoveContainer" containerID="192375eee4b5d01e4694d0ec9847562d172334163d40d6c0f9ff8761ba1fc7a6" Dec 04 12:04:40 crc kubenswrapper[4979]: E1204 12:04:40.681738 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 04 12:04:40 crc kubenswrapper[4979]: E1204 12:04:40.681876 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qp8l4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-2465j_openstack(c0fd1874-37a5-40ab-ab1e-c18a0cffbab6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 12:04:40 crc kubenswrapper[4979]: E1204 12:04:40.683069 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-2465j" podUID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.820208 4979 scope.go:117] "RemoveContainer" containerID="cf9068416f8b059541200f4a2196123cb243bb95381d03f2a7cad3623c7e3f72" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.854523 4979 scope.go:117] "RemoveContainer" containerID="ccf2c599204da74e3bee6b9dfecaa29381899e2aed874f6fbcef0e9e54dc9839" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.891599 4979 scope.go:117] "RemoveContainer" containerID="c8be71aa2ab63e7bba5003b34e684ff345e36633a9b0398fbf6538afd399a7a6" Dec 04 12:04:40 crc kubenswrapper[4979]: I1204 12:04:40.934480 4979 scope.go:117] "RemoveContainer" containerID="194f4362236a1ef75fb2d02b607a54c1b599af7c069ec7c864abb20274f24377" Dec 04 12:04:41 crc kubenswrapper[4979]: W1204 12:04:41.156983 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e3201e7_fdd0_49c2_93bc_96211c54e55a.slice/crio-2f026f1019363b8c0bcdf26af5a74e5da17a69b26391317545212a228bb95b44 WatchSource:0}: Error finding container 2f026f1019363b8c0bcdf26af5a74e5da17a69b26391317545212a228bb95b44: Status 404 returned error can't find the container with id 2f026f1019363b8c0bcdf26af5a74e5da17a69b26391317545212a228bb95b44 Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.157941 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-hfg44"] Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.296437 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.400419 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.532532 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerStarted","Data":"ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439"} Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.535802 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s2g8b" event={"ID":"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112","Type":"ContainerStarted","Data":"a1d5131cd2a4f7ced751fb83f28ba6274321f1df6ac8c22111baa1fa6c0440fc"} Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.543060 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xmtnh" event={"ID":"e9ceee6e-4685-4dc1-8222-57cc6d51a337","Type":"ContainerStarted","Data":"879ae343329fb352b347620cddbfda0a9c4a87f4868acb73890adbe05274d8d3"} Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.545341 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"95a0e5e3-658f-4d68-9387-34966c2a8e1c","Type":"ContainerStarted","Data":"bc4051c92354e8478fa59ca47517a8b38dab985f5003f9ef117ddcbab034a529"} Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.553178 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-s2g8b" podStartSLOduration=2.813894426 podStartE2EDuration="22.553163025s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="2025-12-04 12:04:20.901837439 +0000 UTC m=+1285.176133233" lastFinishedPulling="2025-12-04 12:04:40.641105998 +0000 UTC m=+1304.915401832" observedRunningTime="2025-12-04 12:04:41.551562062 +0000 UTC m=+1305.825857866" watchObservedRunningTime="2025-12-04 12:04:41.553163025 +0000 UTC m=+1305.827458829" Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.558081 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hfg44" event={"ID":"8e3201e7-fdd0-49c2-93bc-96211c54e55a","Type":"ContainerStarted","Data":"1058b8753f8fe274cab6fb08732e2e77aada128e55d54b84e725482bf578e761"} Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.558127 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hfg44" event={"ID":"8e3201e7-fdd0-49c2-93bc-96211c54e55a","Type":"ContainerStarted","Data":"2f026f1019363b8c0bcdf26af5a74e5da17a69b26391317545212a228bb95b44"} Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.566204 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xmtnh" podStartSLOduration=2.722678288 podStartE2EDuration="22.566187409s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="2025-12-04 12:04:20.855582713 +0000 UTC m=+1285.129878517" lastFinishedPulling="2025-12-04 12:04:40.699091834 +0000 UTC m=+1304.973387638" observedRunningTime="2025-12-04 12:04:41.56545641 +0000 UTC m=+1305.839752214" watchObservedRunningTime="2025-12-04 12:04:41.566187409 +0000 UTC m=+1305.840483213" Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.566680 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7","Type":"ContainerStarted","Data":"54d11babcbe0046ee19be2a2ad71fb79be3d97a1b7f1d43e1a47dcc8b2c35871"} Dec 04 12:04:41 crc kubenswrapper[4979]: E1204 12:04:41.568712 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-2465j" podUID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" Dec 04 12:04:41 crc kubenswrapper[4979]: I1204 12:04:41.592067 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-hfg44" podStartSLOduration=9.592045413 podStartE2EDuration="9.592045413s" podCreationTimestamp="2025-12-04 12:04:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:41.583172191 +0000 UTC m=+1305.857468015" watchObservedRunningTime="2025-12-04 12:04:41.592045413 +0000 UTC m=+1305.866341217" Dec 04 12:04:42 crc kubenswrapper[4979]: I1204 12:04:42.209077 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" path="/var/lib/kubelet/pods/333d2328-de05-4d66-a8b1-d0585ddf5f32/volumes" Dec 04 12:04:42 crc kubenswrapper[4979]: I1204 12:04:42.578381 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7","Type":"ContainerStarted","Data":"c29efb06e4e3649086453955662a6ccf8e8a5eb9f46956a25e0d23c4118a24f7"} Dec 04 12:04:42 crc kubenswrapper[4979]: I1204 12:04:42.580885 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerStarted","Data":"12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612"} Dec 04 12:04:42 crc kubenswrapper[4979]: I1204 12:04:42.583241 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"95a0e5e3-658f-4d68-9387-34966c2a8e1c","Type":"ContainerStarted","Data":"df98ae1983f5afe4ff2b92c1de05d7850f1596f0e50d699c019542f0b32b1c05"} Dec 04 12:04:43 crc kubenswrapper[4979]: I1204 12:04:43.190654 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-764c5664d7-rmf45" podUID="333d2328-de05-4d66-a8b1-d0585ddf5f32" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.128:5353: i/o timeout" Dec 04 12:04:43 crc kubenswrapper[4979]: I1204 12:04:43.608670 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7","Type":"ContainerStarted","Data":"78cc8bc9044d8fcae9cd85ae7b724b226af0401ec3dd0144a560f7e93ebe8c61"} Dec 04 12:04:43 crc kubenswrapper[4979]: I1204 12:04:43.628422 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"95a0e5e3-658f-4d68-9387-34966c2a8e1c","Type":"ContainerStarted","Data":"adf32560344263d5e4a29f6de9e9422adb452415702e0da15f1576b2842f89f9"} Dec 04 12:04:43 crc kubenswrapper[4979]: I1204 12:04:43.639308 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.639275925 podStartE2EDuration="4.639275925s" podCreationTimestamp="2025-12-04 12:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:43.632421769 +0000 UTC m=+1307.906717583" watchObservedRunningTime="2025-12-04 12:04:43.639275925 +0000 UTC m=+1307.913571729" Dec 04 12:04:43 crc kubenswrapper[4979]: I1204 12:04:43.658134 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.658116537 podStartE2EDuration="12.658116537s" podCreationTimestamp="2025-12-04 12:04:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:43.655754974 +0000 UTC m=+1307.930050788" watchObservedRunningTime="2025-12-04 12:04:43.658116537 +0000 UTC m=+1307.932412341" Dec 04 12:04:45 crc kubenswrapper[4979]: I1204 12:04:45.648459 4979 generic.go:334] "Generic (PLEG): container finished" podID="8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" containerID="a1d5131cd2a4f7ced751fb83f28ba6274321f1df6ac8c22111baa1fa6c0440fc" exitCode=0 Dec 04 12:04:45 crc kubenswrapper[4979]: I1204 12:04:45.648605 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s2g8b" event={"ID":"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112","Type":"ContainerDied","Data":"a1d5131cd2a4f7ced751fb83f28ba6274321f1df6ac8c22111baa1fa6c0440fc"} Dec 04 12:04:45 crc kubenswrapper[4979]: I1204 12:04:45.653631 4979 generic.go:334] "Generic (PLEG): container finished" podID="8e3201e7-fdd0-49c2-93bc-96211c54e55a" containerID="1058b8753f8fe274cab6fb08732e2e77aada128e55d54b84e725482bf578e761" exitCode=0 Dec 04 12:04:45 crc kubenswrapper[4979]: I1204 12:04:45.653669 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hfg44" event={"ID":"8e3201e7-fdd0-49c2-93bc-96211c54e55a","Type":"ContainerDied","Data":"1058b8753f8fe274cab6fb08732e2e77aada128e55d54b84e725482bf578e761"} Dec 04 12:04:46 crc kubenswrapper[4979]: I1204 12:04:46.690948 4979 generic.go:334] "Generic (PLEG): container finished" podID="e9ceee6e-4685-4dc1-8222-57cc6d51a337" containerID="879ae343329fb352b347620cddbfda0a9c4a87f4868acb73890adbe05274d8d3" exitCode=0 Dec 04 12:04:46 crc kubenswrapper[4979]: I1204 12:04:46.692094 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xmtnh" event={"ID":"e9ceee6e-4685-4dc1-8222-57cc6d51a337","Type":"ContainerDied","Data":"879ae343329fb352b347620cddbfda0a9c4a87f4868acb73890adbe05274d8d3"} Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.184005 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.202696 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.300671 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-scripts\") pod \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.300723 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-scripts\") pod \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.300757 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-fernet-keys\") pod \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.300818 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-combined-ca-bundle\") pod \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.300866 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-config-data\") pod \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.300919 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-credential-keys\") pod \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.301006 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-combined-ca-bundle\") pod \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.301052 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-logs\") pod \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.301092 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfrbc\" (UniqueName: \"kubernetes.io/projected/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-kube-api-access-hfrbc\") pod \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\" (UID: \"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.301140 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zvssz\" (UniqueName: \"kubernetes.io/projected/8e3201e7-fdd0-49c2-93bc-96211c54e55a-kube-api-access-zvssz\") pod \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.301164 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-config-data\") pod \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\" (UID: \"8e3201e7-fdd0-49c2-93bc-96211c54e55a\") " Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.302961 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-logs" (OuterVolumeSpecName: "logs") pod "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" (UID: "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.308130 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "8e3201e7-fdd0-49c2-93bc-96211c54e55a" (UID: "8e3201e7-fdd0-49c2-93bc-96211c54e55a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.308125 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-scripts" (OuterVolumeSpecName: "scripts") pod "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" (UID: "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.308125 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-kube-api-access-hfrbc" (OuterVolumeSpecName: "kube-api-access-hfrbc") pod "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" (UID: "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112"). InnerVolumeSpecName "kube-api-access-hfrbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.308647 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8e3201e7-fdd0-49c2-93bc-96211c54e55a" (UID: "8e3201e7-fdd0-49c2-93bc-96211c54e55a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.309511 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-scripts" (OuterVolumeSpecName: "scripts") pod "8e3201e7-fdd0-49c2-93bc-96211c54e55a" (UID: "8e3201e7-fdd0-49c2-93bc-96211c54e55a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.310678 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e3201e7-fdd0-49c2-93bc-96211c54e55a-kube-api-access-zvssz" (OuterVolumeSpecName: "kube-api-access-zvssz") pod "8e3201e7-fdd0-49c2-93bc-96211c54e55a" (UID: "8e3201e7-fdd0-49c2-93bc-96211c54e55a"). InnerVolumeSpecName "kube-api-access-zvssz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.335035 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e3201e7-fdd0-49c2-93bc-96211c54e55a" (UID: "8e3201e7-fdd0-49c2-93bc-96211c54e55a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.335686 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" (UID: "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.350977 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-config-data" (OuterVolumeSpecName: "config-data") pod "8e3201e7-fdd0-49c2-93bc-96211c54e55a" (UID: "8e3201e7-fdd0-49c2-93bc-96211c54e55a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.351592 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-config-data" (OuterVolumeSpecName: "config-data") pod "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" (UID: "8bc5e0d4-e60f-465c-8ed5-c9a4203f4112"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403235 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403265 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403275 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfrbc\" (UniqueName: \"kubernetes.io/projected/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-kube-api-access-hfrbc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403286 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zvssz\" (UniqueName: \"kubernetes.io/projected/8e3201e7-fdd0-49c2-93bc-96211c54e55a-kube-api-access-zvssz\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403294 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403313 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403320 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403328 4979 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403336 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403344 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.403352 4979 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/8e3201e7-fdd0-49c2-93bc-96211c54e55a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.730496 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-hfg44" event={"ID":"8e3201e7-fdd0-49c2-93bc-96211c54e55a","Type":"ContainerDied","Data":"2f026f1019363b8c0bcdf26af5a74e5da17a69b26391317545212a228bb95b44"} Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.730551 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f026f1019363b8c0bcdf26af5a74e5da17a69b26391317545212a228bb95b44" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.730763 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-hfg44" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.750757 4979 generic.go:334] "Generic (PLEG): container finished" podID="257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" containerID="aed86b6e4e107649f8fb95f66dcae276fcc47522a30318a9927a27f1a5c5a593" exitCode=0 Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.750838 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-scv9t" event={"ID":"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9","Type":"ContainerDied","Data":"aed86b6e4e107649f8fb95f66dcae276fcc47522a30318a9927a27f1a5c5a593"} Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.753719 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerStarted","Data":"51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25"} Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.759826 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-s2g8b" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.760041 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-s2g8b" event={"ID":"8bc5e0d4-e60f-465c-8ed5-c9a4203f4112","Type":"ContainerDied","Data":"d3cc2b03b2165a86339a57d1724f4f0b1b40c9fe4538b5262557bd24b0d85e71"} Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.760115 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3cc2b03b2165a86339a57d1724f4f0b1b40c9fe4538b5262557bd24b0d85e71" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.813870 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5667db5f8d-q4ffn"] Dec 04 12:04:47 crc kubenswrapper[4979]: E1204 12:04:47.814235 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" containerName="placement-db-sync" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.814257 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" containerName="placement-db-sync" Dec 04 12:04:47 crc kubenswrapper[4979]: E1204 12:04:47.814276 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e3201e7-fdd0-49c2-93bc-96211c54e55a" containerName="keystone-bootstrap" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.814282 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e3201e7-fdd0-49c2-93bc-96211c54e55a" containerName="keystone-bootstrap" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.814490 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" containerName="placement-db-sync" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.814519 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e3201e7-fdd0-49c2-93bc-96211c54e55a" containerName="keystone-bootstrap" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.815355 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.820680 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.820940 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.821256 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.821383 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bbh4d" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.821484 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.827539 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5667db5f8d-q4ffn"] Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.909195 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7c66cf746c-gkxxt"] Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.913487 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916504 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-config-data\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916564 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-combined-ca-bundle\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916610 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-logs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916633 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-scripts\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916647 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-public-tls-certs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916687 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-internal-tls-certs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.916734 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zlfl\" (UniqueName: \"kubernetes.io/projected/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-kube-api-access-2zlfl\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.918493 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.918525 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.918540 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.918679 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-zbvzn" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.918846 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.918896 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 12:04:47 crc kubenswrapper[4979]: I1204 12:04:47.928312 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c66cf746c-gkxxt"] Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.017826 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-scripts\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.017874 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-logs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.017900 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-scripts\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.017918 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-public-tls-certs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018034 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-config-data\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018100 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sk8k\" (UniqueName: \"kubernetes.io/projected/983663bf-7439-4648-8237-1174fe609970-kube-api-access-6sk8k\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-internal-tls-certs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018225 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-combined-ca-bundle\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018807 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-public-tls-certs\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018631 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-logs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.018860 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-fernet-keys\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.019041 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zlfl\" (UniqueName: \"kubernetes.io/projected/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-kube-api-access-2zlfl\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.019122 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-credential-keys\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.019169 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-config-data\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.019250 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-combined-ca-bundle\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.019338 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-internal-tls-certs\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.022189 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-public-tls-certs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.034097 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-config-data\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.034130 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-scripts\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.034137 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-internal-tls-certs\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.036990 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-combined-ca-bundle\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.038462 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zlfl\" (UniqueName: \"kubernetes.io/projected/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-kube-api-access-2zlfl\") pod \"placement-5667db5f8d-q4ffn\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121413 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-scripts\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121498 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-config-data\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121522 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sk8k\" (UniqueName: \"kubernetes.io/projected/983663bf-7439-4648-8237-1174fe609970-kube-api-access-6sk8k\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121572 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-combined-ca-bundle\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121610 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-public-tls-certs\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121639 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-fernet-keys\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121680 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-credential-keys\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.121737 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-internal-tls-certs\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.126521 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-config-data\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.127453 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-fernet-keys\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.127492 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-scripts\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.127548 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-internal-tls-certs\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.127587 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-public-tls-certs\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.127754 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-credential-keys\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.128996 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-combined-ca-bundle\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.136871 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sk8k\" (UniqueName: \"kubernetes.io/projected/983663bf-7439-4648-8237-1174fe609970-kube-api-access-6sk8k\") pod \"keystone-7c66cf746c-gkxxt\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.181407 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.192749 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.240741 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.324280 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckxtc\" (UniqueName: \"kubernetes.io/projected/e9ceee6e-4685-4dc1-8222-57cc6d51a337-kube-api-access-ckxtc\") pod \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.324404 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-combined-ca-bundle\") pod \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.324538 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-db-sync-config-data\") pod \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\" (UID: \"e9ceee6e-4685-4dc1-8222-57cc6d51a337\") " Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.329481 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "e9ceee6e-4685-4dc1-8222-57cc6d51a337" (UID: "e9ceee6e-4685-4dc1-8222-57cc6d51a337"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.329651 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9ceee6e-4685-4dc1-8222-57cc6d51a337-kube-api-access-ckxtc" (OuterVolumeSpecName: "kube-api-access-ckxtc") pod "e9ceee6e-4685-4dc1-8222-57cc6d51a337" (UID: "e9ceee6e-4685-4dc1-8222-57cc6d51a337"). InnerVolumeSpecName "kube-api-access-ckxtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.349955 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9ceee6e-4685-4dc1-8222-57cc6d51a337" (UID: "e9ceee6e-4685-4dc1-8222-57cc6d51a337"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.430585 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckxtc\" (UniqueName: \"kubernetes.io/projected/e9ceee6e-4685-4dc1-8222-57cc6d51a337-kube-api-access-ckxtc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.430682 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.430698 4979 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/e9ceee6e-4685-4dc1-8222-57cc6d51a337-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.662039 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5667db5f8d-q4ffn"] Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.769036 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xmtnh" event={"ID":"e9ceee6e-4685-4dc1-8222-57cc6d51a337","Type":"ContainerDied","Data":"41686986fde3ea66945072a0b937d7b775a014661a9f06b74fa339a072a4c98d"} Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.769954 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41686986fde3ea66945072a0b937d7b775a014661a9f06b74fa339a072a4c98d" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.769487 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xmtnh" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.770650 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5667db5f8d-q4ffn" event={"ID":"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd","Type":"ContainerStarted","Data":"43712bce63202df71bb07043fdbae8edb9815e8cab5ccc23b36112d41f7aa833"} Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.782828 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7c66cf746c-gkxxt"] Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.955038 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6bf4748b96-fzr6d"] Dec 04 12:04:48 crc kubenswrapper[4979]: E1204 12:04:48.955810 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9ceee6e-4685-4dc1-8222-57cc6d51a337" containerName="barbican-db-sync" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.955825 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9ceee6e-4685-4dc1-8222-57cc6d51a337" containerName="barbican-db-sync" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.956001 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9ceee6e-4685-4dc1-8222-57cc6d51a337" containerName="barbican-db-sync" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.956891 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.964429 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.964661 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.964923 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-zgx4w" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.966380 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7b84847487-r8c7n"] Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.968638 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.975695 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.979275 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b84847487-r8c7n"] Dec 04 12:04:48 crc kubenswrapper[4979]: I1204 12:04:48.991280 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6bf4748b96-fzr6d"] Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046429 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a1863-ce1d-4acd-9d50-6987b60e05f0-logs\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046487 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046505 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/869a1863-ce1d-4acd-9d50-6987b60e05f0-kube-api-access-9kbvg\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046543 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data-custom\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2952g\" (UniqueName: \"kubernetes.io/projected/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-kube-api-access-2952g\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046583 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data-custom\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046637 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-combined-ca-bundle\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046664 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-logs\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046703 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-combined-ca-bundle\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.046725 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.136624 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-82jz8"] Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.138007 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148109 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-combined-ca-bundle\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148159 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-logs\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148207 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-combined-ca-bundle\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148258 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a1863-ce1d-4acd-9d50-6987b60e05f0-logs\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148275 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148291 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/869a1863-ce1d-4acd-9d50-6987b60e05f0-kube-api-access-9kbvg\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data-custom\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148353 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2952g\" (UniqueName: \"kubernetes.io/projected/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-kube-api-access-2952g\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.148374 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data-custom\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.149592 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-logs\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.156266 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-82jz8"] Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.156705 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a1863-ce1d-4acd-9d50-6987b60e05f0-logs\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.172240 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-combined-ca-bundle\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.175310 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data-custom\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.176111 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-combined-ca-bundle\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.176973 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data-custom\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.177251 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/869a1863-ce1d-4acd-9d50-6987b60e05f0-kube-api-access-9kbvg\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.178221 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.185263 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2952g\" (UniqueName: \"kubernetes.io/projected/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-kube-api-access-2952g\") pod \"barbican-keystone-listener-6bf4748b96-fzr6d\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.197984 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data\") pod \"barbican-worker-7b84847487-r8c7n\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.250759 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.250799 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.250881 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-config\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.250898 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.250942 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwvb5\" (UniqueName: \"kubernetes.io/projected/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-kube-api-access-dwvb5\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.250981 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.307222 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.313714 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58f74675fd-6lkn6"] Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.320004 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.322817 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58f74675fd-6lkn6"] Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.325120 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.339735 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.353626 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.353681 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.353760 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-config\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.353786 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.353830 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwvb5\" (UniqueName: \"kubernetes.io/projected/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-kube-api-access-dwvb5\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.353874 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.354599 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.354643 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-config\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.354881 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.355214 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.355261 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.380353 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.394184 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwvb5\" (UniqueName: \"kubernetes.io/projected/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-kube-api-access-dwvb5\") pod \"dnsmasq-dns-586bdc5f9-82jz8\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458213 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-combined-ca-bundle\") pod \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458352 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-config\") pod \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458448 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-549dz\" (UniqueName: \"kubernetes.io/projected/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-kube-api-access-549dz\") pod \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\" (UID: \"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9\") " Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458662 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dee1bb55-edbe-4cc0-989d-918bbd9dd487-logs\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458686 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data-custom\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458712 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458771 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-combined-ca-bundle\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.458817 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgnxh\" (UniqueName: \"kubernetes.io/projected/dee1bb55-edbe-4cc0-989d-918bbd9dd487-kube-api-access-dgnxh\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.472924 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-kube-api-access-549dz" (OuterVolumeSpecName: "kube-api-access-549dz") pod "257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" (UID: "257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9"). InnerVolumeSpecName "kube-api-access-549dz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.483983 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-config" (OuterVolumeSpecName: "config") pod "257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" (UID: "257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.507350 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" (UID: "257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.537980 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561207 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dee1bb55-edbe-4cc0-989d-918bbd9dd487-logs\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561273 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data-custom\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561324 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561407 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-combined-ca-bundle\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561459 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgnxh\" (UniqueName: \"kubernetes.io/projected/dee1bb55-edbe-4cc0-989d-918bbd9dd487-kube-api-access-dgnxh\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561558 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561574 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.561585 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-549dz\" (UniqueName: \"kubernetes.io/projected/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9-kube-api-access-549dz\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.562558 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dee1bb55-edbe-4cc0-989d-918bbd9dd487-logs\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.566883 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data-custom\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.567289 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-combined-ca-bundle\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.567784 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.587121 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgnxh\" (UniqueName: \"kubernetes.io/projected/dee1bb55-edbe-4cc0-989d-918bbd9dd487-kube-api-access-dgnxh\") pod \"barbican-api-58f74675fd-6lkn6\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.624714 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6bf4748b96-fzr6d"] Dec 04 12:04:49 crc kubenswrapper[4979]: W1204 12:04:49.627949 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8589c26_b1f1_48b2_a0cd_5b1bdc62336a.slice/crio-e9e0d76af00c43a82906c62c6a0428bf37ccb44e1b696e17730d74ec00a096d2 WatchSource:0}: Error finding container e9e0d76af00c43a82906c62c6a0428bf37ccb44e1b696e17730d74ec00a096d2: Status 404 returned error can't find the container with id e9e0d76af00c43a82906c62c6a0428bf37ccb44e1b696e17730d74ec00a096d2 Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.645800 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.810889 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c66cf746c-gkxxt" event={"ID":"983663bf-7439-4648-8237-1174fe609970","Type":"ContainerStarted","Data":"91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1"} Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.810997 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.811010 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c66cf746c-gkxxt" event={"ID":"983663bf-7439-4648-8237-1174fe609970","Type":"ContainerStarted","Data":"6913d7234ddba603f928d309b2145b558a0fdda4ad8c35b0e9100d92e34e5a13"} Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.824660 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-scv9t" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.824673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-scv9t" event={"ID":"257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9","Type":"ContainerDied","Data":"35478a4648d79d8ec051ef0badd4a1cca8a17a5324912142ebd7152609afd2f1"} Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.825440 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35478a4648d79d8ec051ef0badd4a1cca8a17a5324912142ebd7152609afd2f1" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.827499 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5667db5f8d-q4ffn" event={"ID":"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd","Type":"ContainerStarted","Data":"89858cce1ad717d7a0855a8184b9e88c16893052de81ca4a07efe6a69e691109"} Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.840607 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" event={"ID":"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a","Type":"ContainerStarted","Data":"e9e0d76af00c43a82906c62c6a0428bf37ccb44e1b696e17730d74ec00a096d2"} Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.845368 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7c66cf746c-gkxxt" podStartSLOduration=2.845350634 podStartE2EDuration="2.845350634s" podCreationTimestamp="2025-12-04 12:04:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:49.838525219 +0000 UTC m=+1314.112821023" watchObservedRunningTime="2025-12-04 12:04:49.845350634 +0000 UTC m=+1314.119646438" Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.936250 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7b84847487-r8c7n"] Dec 04 12:04:49 crc kubenswrapper[4979]: I1204 12:04:49.983652 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-82jz8"] Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.028202 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qt9j9"] Dec 04 12:04:50 crc kubenswrapper[4979]: E1204 12:04:50.031789 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" containerName="neutron-db-sync" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.031816 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" containerName="neutron-db-sync" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.031993 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" containerName="neutron-db-sync" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.034596 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.051167 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qt9j9"] Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.086189 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5cfc776c66-nvfsf"] Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.088707 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.090012 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-jltrn" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.090587 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.090741 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.091425 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.115175 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cfc776c66-nvfsf"] Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.122968 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-82jz8"] Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178118 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-config\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178173 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-svc\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178195 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178215 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpz8p\" (UniqueName: \"kubernetes.io/projected/0c55939b-bdfd-4deb-8d49-11036a75f364-kube-api-access-cpz8p\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178254 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-ovndb-tls-certs\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178277 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178319 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-config\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178370 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178390 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x784c\" (UniqueName: \"kubernetes.io/projected/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-kube-api-access-x784c\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178443 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-combined-ca-bundle\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.178463 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-httpd-config\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.282256 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-ovndb-tls-certs\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.282932 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.283934 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-config\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.284227 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.284452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x784c\" (UniqueName: \"kubernetes.io/projected/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-kube-api-access-x784c\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.285470 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-config\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.287502 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.287649 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.288724 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-combined-ca-bundle\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.288830 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-httpd-config\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.289067 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-config\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.289184 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-svc\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.289269 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.289357 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpz8p\" (UniqueName: \"kubernetes.io/projected/0c55939b-bdfd-4deb-8d49-11036a75f364-kube-api-access-cpz8p\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.290469 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-ovndb-tls-certs\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.292419 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-svc\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.292979 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.295433 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-config\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.299054 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-httpd-config\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.309790 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-combined-ca-bundle\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.313163 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x784c\" (UniqueName: \"kubernetes.io/projected/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-kube-api-access-x784c\") pod \"neutron-5cfc776c66-nvfsf\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.327520 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpz8p\" (UniqueName: \"kubernetes.io/projected/0c55939b-bdfd-4deb-8d49-11036a75f364-kube-api-access-cpz8p\") pod \"dnsmasq-dns-85ff748b95-qt9j9\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.376201 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.402692 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58f74675fd-6lkn6"] Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.436069 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.494673 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.494841 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.532107 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.616657 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.856872 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58f74675fd-6lkn6" event={"ID":"dee1bb55-edbe-4cc0-989d-918bbd9dd487","Type":"ContainerStarted","Data":"cffdab1c4f49fdc4bf399e736e5dc7c1dfdd72058185891a8eb144e96d678108"} Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.861013 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" event={"ID":"d96b7b5f-49c1-4ebe-80d2-e47b20f57521","Type":"ContainerStarted","Data":"1f06035991ecb1d210a67cf1e7fdfecb364783fe91ddbe5c3b1428f8ba0e4237"} Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.861040 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" event={"ID":"d96b7b5f-49c1-4ebe-80d2-e47b20f57521","Type":"ContainerStarted","Data":"009b0352852c877901b77027034d3eb6aa7188a94030b1b29abe81038328000a"} Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.863042 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5667db5f8d-q4ffn" event={"ID":"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd","Type":"ContainerStarted","Data":"dfee452098fffdbca65d5be95466eeb6448db8377ba02e117a1cd9cea7962b7f"} Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.863412 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.863500 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.865730 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b84847487-r8c7n" event={"ID":"869a1863-ce1d-4acd-9d50-6987b60e05f0","Type":"ContainerStarted","Data":"209376b52f61c881989d79e3c6add103d70f6c95015802fac48f8e7f2772b22b"} Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.866437 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.866487 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 12:04:50 crc kubenswrapper[4979]: I1204 12:04:50.888717 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5667db5f8d-q4ffn" podStartSLOduration=3.8887000929999997 podStartE2EDuration="3.888700093s" podCreationTimestamp="2025-12-04 12:04:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:50.882237417 +0000 UTC m=+1315.156533241" watchObservedRunningTime="2025-12-04 12:04:50.888700093 +0000 UTC m=+1315.162995897" Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.004412 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cfc776c66-nvfsf"] Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.029806 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qt9j9"] Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.907698 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" event={"ID":"0c55939b-bdfd-4deb-8d49-11036a75f364","Type":"ContainerStarted","Data":"3a3bf5fbfe356969e1159ee081d25588b6129f114e251015092d664cc43f7b8d"} Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.912872 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cfc776c66-nvfsf" event={"ID":"30547b8b-d0c9-4473-9f34-f8b9bca3e20a","Type":"ContainerStarted","Data":"5990fa32fb947afa7d83789dd1952f5630ba7467ccb7dcc3034d77d271c46919"} Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.916117 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58f74675fd-6lkn6" event={"ID":"dee1bb55-edbe-4cc0-989d-918bbd9dd487","Type":"ContainerStarted","Data":"a69924e3d2ab794ab8931fce1d8a12a718cbc7461f1b93f1393b6e1bcfd0c441"} Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.917935 4979 generic.go:334] "Generic (PLEG): container finished" podID="d96b7b5f-49c1-4ebe-80d2-e47b20f57521" containerID="1f06035991ecb1d210a67cf1e7fdfecb364783fe91ddbe5c3b1428f8ba0e4237" exitCode=0 Dec 04 12:04:51 crc kubenswrapper[4979]: I1204 12:04:51.918228 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" event={"ID":"d96b7b5f-49c1-4ebe-80d2-e47b20f57521","Type":"ContainerDied","Data":"1f06035991ecb1d210a67cf1e7fdfecb364783fe91ddbe5c3b1428f8ba0e4237"} Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.103942 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.103999 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.143969 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.163020 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.202673 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.249641 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-nb\") pod \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.249726 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-sb\") pod \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.249773 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-swift-storage-0\") pod \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.249830 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-svc\") pod \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.249933 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-config\") pod \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.250047 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwvb5\" (UniqueName: \"kubernetes.io/projected/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-kube-api-access-dwvb5\") pod \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\" (UID: \"d96b7b5f-49c1-4ebe-80d2-e47b20f57521\") " Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.258485 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-kube-api-access-dwvb5" (OuterVolumeSpecName: "kube-api-access-dwvb5") pod "d96b7b5f-49c1-4ebe-80d2-e47b20f57521" (UID: "d96b7b5f-49c1-4ebe-80d2-e47b20f57521"). InnerVolumeSpecName "kube-api-access-dwvb5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.276813 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d96b7b5f-49c1-4ebe-80d2-e47b20f57521" (UID: "d96b7b5f-49c1-4ebe-80d2-e47b20f57521"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.281457 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d96b7b5f-49c1-4ebe-80d2-e47b20f57521" (UID: "d96b7b5f-49c1-4ebe-80d2-e47b20f57521"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.282419 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d96b7b5f-49c1-4ebe-80d2-e47b20f57521" (UID: "d96b7b5f-49c1-4ebe-80d2-e47b20f57521"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.288152 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-config" (OuterVolumeSpecName: "config") pod "d96b7b5f-49c1-4ebe-80d2-e47b20f57521" (UID: "d96b7b5f-49c1-4ebe-80d2-e47b20f57521"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.289224 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d96b7b5f-49c1-4ebe-80d2-e47b20f57521" (UID: "d96b7b5f-49c1-4ebe-80d2-e47b20f57521"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.352717 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.352748 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.352759 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.352767 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwvb5\" (UniqueName: \"kubernetes.io/projected/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-kube-api-access-dwvb5\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.352776 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.352784 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d96b7b5f-49c1-4ebe-80d2-e47b20f57521-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.964924 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58f74675fd-6lkn6" event={"ID":"dee1bb55-edbe-4cc0-989d-918bbd9dd487","Type":"ContainerStarted","Data":"11b2fee3fbc24534e5dea3a5dc9064a3b8e00d8dec1712ba02e67ce3f008705b"} Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.965423 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.965493 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.986163 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.987859 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58f74675fd-6lkn6" podStartSLOduration=3.987841387 podStartE2EDuration="3.987841387s" podCreationTimestamp="2025-12-04 12:04:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:04:52.987795036 +0000 UTC m=+1317.262090850" watchObservedRunningTime="2025-12-04 12:04:52.987841387 +0000 UTC m=+1317.262137191" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.988132 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-82jz8" event={"ID":"d96b7b5f-49c1-4ebe-80d2-e47b20f57521","Type":"ContainerDied","Data":"009b0352852c877901b77027034d3eb6aa7188a94030b1b29abe81038328000a"} Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.988186 4979 scope.go:117] "RemoveContainer" containerID="1f06035991ecb1d210a67cf1e7fdfecb364783fe91ddbe5c3b1428f8ba0e4237" Dec 04 12:04:52 crc kubenswrapper[4979]: I1204 12:04:52.997603 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b84847487-r8c7n" event={"ID":"869a1863-ce1d-4acd-9d50-6987b60e05f0","Type":"ContainerStarted","Data":"de14bd59b0b06af679f770868d3c98798bbffcb02c92b7336ab4b61c2bc90a5b"} Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.008553 4979 generic.go:334] "Generic (PLEG): container finished" podID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerID="75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c" exitCode=0 Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.008673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" event={"ID":"0c55939b-bdfd-4deb-8d49-11036a75f364","Type":"ContainerDied","Data":"75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c"} Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.015265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cfc776c66-nvfsf" event={"ID":"30547b8b-d0c9-4473-9f34-f8b9bca3e20a","Type":"ContainerStarted","Data":"2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b"} Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.032120 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" event={"ID":"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a","Type":"ContainerStarted","Data":"4a64fd8d676b9faa863cf1c62e916da023a1db555a91558e98ac315dc5777bab"} Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.032171 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.032228 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.032238 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.032903 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.084349 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-82jz8"] Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.118629 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-82jz8"] Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.401709 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 12:04:53 crc kubenswrapper[4979]: I1204 12:04:53.403754 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.127508 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-698495cf57-xzqk7"] Dec 04 12:04:54 crc kubenswrapper[4979]: E1204 12:04:54.128282 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d96b7b5f-49c1-4ebe-80d2-e47b20f57521" containerName="init" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.128314 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d96b7b5f-49c1-4ebe-80d2-e47b20f57521" containerName="init" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.128533 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d96b7b5f-49c1-4ebe-80d2-e47b20f57521" containerName="init" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.131149 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.133934 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.134569 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.138694 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-698495cf57-xzqk7"] Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.211909 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d96b7b5f-49c1-4ebe-80d2-e47b20f57521" path="/var/lib/kubelet/pods/d96b7b5f-49c1-4ebe-80d2-e47b20f57521/volumes" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286268 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-combined-ca-bundle\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286408 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-public-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286457 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-config\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286489 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-ovndb-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286602 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-httpd-config\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286717 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xthth\" (UniqueName: \"kubernetes.io/projected/d5d21f75-aea0-4306-843d-c3eb782cdd78-kube-api-access-xthth\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.286770 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-internal-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388125 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xthth\" (UniqueName: \"kubernetes.io/projected/d5d21f75-aea0-4306-843d-c3eb782cdd78-kube-api-access-xthth\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388213 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-internal-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388245 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-combined-ca-bundle\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388449 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-public-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388529 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-config\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-ovndb-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.388709 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-httpd-config\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.394503 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-httpd-config\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.394894 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-internal-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.395610 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-config\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.396643 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-ovndb-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.403956 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-combined-ca-bundle\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.411097 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xthth\" (UniqueName: \"kubernetes.io/projected/d5d21f75-aea0-4306-843d-c3eb782cdd78-kube-api-access-xthth\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.411542 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-public-tls-certs\") pod \"neutron-698495cf57-xzqk7\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:54 crc kubenswrapper[4979]: I1204 12:04:54.456840 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.048936 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.048969 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.669479 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.902564 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-57cfcdd946-gb75v"] Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.904415 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.914527 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.914813 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 04 12:04:55 crc kubenswrapper[4979]: I1204 12:04:55.976357 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57cfcdd946-gb75v"] Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021216 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021284 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-internal-tls-certs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021324 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-public-tls-certs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021371 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data-custom\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021391 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-combined-ca-bundle\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021424 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ba50945-6ce9-4a71-9f58-e436b646c779-logs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.021462 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqtd4\" (UniqueName: \"kubernetes.io/projected/1ba50945-6ce9-4a71-9f58-e436b646c779-kube-api-access-gqtd4\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.022457 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-internal-tls-certs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123455 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-public-tls-certs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123530 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data-custom\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123555 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-combined-ca-bundle\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123599 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ba50945-6ce9-4a71-9f58-e436b646c779-logs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.123681 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqtd4\" (UniqueName: \"kubernetes.io/projected/1ba50945-6ce9-4a71-9f58-e436b646c779-kube-api-access-gqtd4\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.124594 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ba50945-6ce9-4a71-9f58-e436b646c779-logs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.129224 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.130208 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data-custom\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.133817 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-internal-tls-certs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.133930 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-public-tls-certs\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.143184 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-combined-ca-bundle\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.143842 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqtd4\" (UniqueName: \"kubernetes.io/projected/1ba50945-6ce9-4a71-9f58-e436b646c779-kube-api-access-gqtd4\") pod \"barbican-api-57cfcdd946-gb75v\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:56 crc kubenswrapper[4979]: I1204 12:04:56.247793 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:04:58 crc kubenswrapper[4979]: I1204 12:04:58.041077 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:04:58 crc kubenswrapper[4979]: I1204 12:04:58.041433 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:05:01 crc kubenswrapper[4979]: I1204 12:05:01.152931 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:05:01 crc kubenswrapper[4979]: I1204 12:05:01.216730 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:05:01 crc kubenswrapper[4979]: I1204 12:05:01.491314 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-57cfcdd946-gb75v"] Dec 04 12:05:01 crc kubenswrapper[4979]: I1204 12:05:01.546782 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-698495cf57-xzqk7"] Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.126654 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cfcdd946-gb75v" event={"ID":"1ba50945-6ce9-4a71-9f58-e436b646c779","Type":"ContainerStarted","Data":"f79822b9ecbd0ae0552d4c0bef356ac9d8b4b613b0aa7af9b7905e5627f9d542"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.126995 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cfcdd946-gb75v" event={"ID":"1ba50945-6ce9-4a71-9f58-e436b646c779","Type":"ContainerStarted","Data":"7852e4534fa46cd05339048e17823fa84f59d75ecdf087a64e4cae75ed60dd30"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.131174 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerStarted","Data":"968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.131503 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-central-agent" containerID="cri-o://ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439" gracePeriod=30 Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.131816 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.132172 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="proxy-httpd" containerID="cri-o://968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa" gracePeriod=30 Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.132234 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="sg-core" containerID="cri-o://51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25" gracePeriod=30 Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.132286 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-notification-agent" containerID="cri-o://12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612" gracePeriod=30 Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.169749 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.637728301 podStartE2EDuration="43.169727626s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="2025-12-04 12:04:20.733862465 +0000 UTC m=+1285.008158269" lastFinishedPulling="2025-12-04 12:05:01.26586179 +0000 UTC m=+1325.540157594" observedRunningTime="2025-12-04 12:05:02.153377051 +0000 UTC m=+1326.427672855" watchObservedRunningTime="2025-12-04 12:05:02.169727626 +0000 UTC m=+1326.444023430" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.172218 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" event={"ID":"0c55939b-bdfd-4deb-8d49-11036a75f364","Type":"ContainerStarted","Data":"b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.174383 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.188006 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cfc776c66-nvfsf" event={"ID":"30547b8b-d0c9-4473-9f34-f8b9bca3e20a","Type":"ContainerStarted","Data":"384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.188270 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.204732 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" podStartSLOduration=13.204324646 podStartE2EDuration="13.204324646s" podCreationTimestamp="2025-12-04 12:04:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:02.193994516 +0000 UTC m=+1326.468290330" watchObservedRunningTime="2025-12-04 12:05:02.204324646 +0000 UTC m=+1326.478620460" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.215788 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" event={"ID":"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a","Type":"ContainerStarted","Data":"b68d477aa81babb95a8e72c26e66dc3b949d75be0432ca0d51c1a917acf6f6e3"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.218354 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b84847487-r8c7n" event={"ID":"869a1863-ce1d-4acd-9d50-6987b60e05f0","Type":"ContainerStarted","Data":"15976ba28c5c9feffdf2219916bb384ba02e85b8e1455eb217e64c14c2a43e76"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.220954 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5cfc776c66-nvfsf" podStartSLOduration=12.220930018 podStartE2EDuration="12.220930018s" podCreationTimestamp="2025-12-04 12:04:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:02.214250046 +0000 UTC m=+1326.488545880" watchObservedRunningTime="2025-12-04 12:05:02.220930018 +0000 UTC m=+1326.495225832" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.236074 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698495cf57-xzqk7" event={"ID":"d5d21f75-aea0-4306-843d-c3eb782cdd78","Type":"ContainerStarted","Data":"225feb849cbf0c27ccb08142f91bf2d2a45a52845a9b008d7dd3f641eb0c0175"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.236129 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698495cf57-xzqk7" event={"ID":"d5d21f75-aea0-4306-843d-c3eb782cdd78","Type":"ContainerStarted","Data":"3d5f1ea774048bd6ce4fe50f30d87ddad110465fde9f7159b9ce3766eddadf93"} Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.236375 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.237874 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" podStartSLOduration=12.092265731 podStartE2EDuration="14.237858868s" podCreationTimestamp="2025-12-04 12:04:48 +0000 UTC" firstStartedPulling="2025-12-04 12:04:49.630733039 +0000 UTC m=+1313.905028843" lastFinishedPulling="2025-12-04 12:04:51.776326176 +0000 UTC m=+1316.050621980" observedRunningTime="2025-12-04 12:05:02.232728739 +0000 UTC m=+1326.507024543" watchObservedRunningTime="2025-12-04 12:05:02.237858868 +0000 UTC m=+1326.512154672" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.266001 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7b84847487-r8c7n" podStartSLOduration=12.42878427 podStartE2EDuration="14.265983122s" podCreationTimestamp="2025-12-04 12:04:48 +0000 UTC" firstStartedPulling="2025-12-04 12:04:49.969434468 +0000 UTC m=+1314.243730272" lastFinishedPulling="2025-12-04 12:04:51.80663332 +0000 UTC m=+1316.080929124" observedRunningTime="2025-12-04 12:05:02.249016371 +0000 UTC m=+1326.523312175" watchObservedRunningTime="2025-12-04 12:05:02.265983122 +0000 UTC m=+1326.540278926" Dec 04 12:05:02 crc kubenswrapper[4979]: I1204 12:05:02.280648 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-698495cf57-xzqk7" podStartSLOduration=8.280620871 podStartE2EDuration="8.280620871s" podCreationTimestamp="2025-12-04 12:04:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:02.269873898 +0000 UTC m=+1326.544169702" watchObservedRunningTime="2025-12-04 12:05:02.280620871 +0000 UTC m=+1326.554916675" Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.249449 4979 generic.go:334] "Generic (PLEG): container finished" podID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerID="968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa" exitCode=0 Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.249739 4979 generic.go:334] "Generic (PLEG): container finished" podID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerID="51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25" exitCode=2 Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.249750 4979 generic.go:334] "Generic (PLEG): container finished" podID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerID="ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439" exitCode=0 Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.249606 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerDied","Data":"968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa"} Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.249820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerDied","Data":"51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25"} Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.249836 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerDied","Data":"ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439"} Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.251749 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698495cf57-xzqk7" event={"ID":"d5d21f75-aea0-4306-843d-c3eb782cdd78","Type":"ContainerStarted","Data":"4e4c287b0e57b02ab10e8858559ba9bd0849ef197cf2a81904670a23b3ad447d"} Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.254628 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2465j" event={"ID":"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6","Type":"ContainerStarted","Data":"07755584661f943a7836f5d09c354a242f8dc548a3c61640b8eac7753c254ff9"} Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.257963 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cfcdd946-gb75v" event={"ID":"1ba50945-6ce9-4a71-9f58-e436b646c779","Type":"ContainerStarted","Data":"c1697a681e440b41b665f6d4ba7c31e3a29aecbb77162bfc4b85770584ea8a4e"} Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.308911 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-57cfcdd946-gb75v" podStartSLOduration=8.308889838 podStartE2EDuration="8.308889838s" podCreationTimestamp="2025-12-04 12:04:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:03.297654233 +0000 UTC m=+1327.571950037" watchObservedRunningTime="2025-12-04 12:05:03.308889838 +0000 UTC m=+1327.583185642" Dec 04 12:05:03 crc kubenswrapper[4979]: I1204 12:05:03.311092 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-2465j" podStartSLOduration=3.900333047 podStartE2EDuration="44.311078298s" podCreationTimestamp="2025-12-04 12:04:19 +0000 UTC" firstStartedPulling="2025-12-04 12:04:20.856321082 +0000 UTC m=+1285.130616886" lastFinishedPulling="2025-12-04 12:05:01.267066333 +0000 UTC m=+1325.541362137" observedRunningTime="2025-12-04 12:05:03.281324629 +0000 UTC m=+1327.555620433" watchObservedRunningTime="2025-12-04 12:05:03.311078298 +0000 UTC m=+1327.585374122" Dec 04 12:05:04 crc kubenswrapper[4979]: I1204 12:05:04.267200 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:05:04 crc kubenswrapper[4979]: I1204 12:05:04.267622 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.262046 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.280625 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerDied","Data":"12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612"} Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.280643 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.280673 4979 scope.go:117] "RemoveContainer" containerID="968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.280585 4979 generic.go:334] "Generic (PLEG): container finished" podID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerID="12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612" exitCode=0 Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.280906 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed","Type":"ContainerDied","Data":"fa83fc2b3c315f66cb7005de5f65fd09170a5b16c1264ccd42fa5c54bf025c0d"} Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289715 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-scripts\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289756 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-sg-core-conf-yaml\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289803 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-run-httpd\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289861 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-config-data\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289887 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-combined-ca-bundle\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289932 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxbl5\" (UniqueName: \"kubernetes.io/projected/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-kube-api-access-bxbl5\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.289961 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-log-httpd\") pod \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\" (UID: \"f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed\") " Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.293145 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.293472 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.297227 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-scripts" (OuterVolumeSpecName: "scripts") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.297420 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-kube-api-access-bxbl5" (OuterVolumeSpecName: "kube-api-access-bxbl5") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "kube-api-access-bxbl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.312631 4979 scope.go:117] "RemoveContainer" containerID="51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.323794 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.362534 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.391449 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.391494 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxbl5\" (UniqueName: \"kubernetes.io/projected/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-kube-api-access-bxbl5\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.391505 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.391512 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.391522 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.391530 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.393763 4979 scope.go:117] "RemoveContainer" containerID="12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.408597 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-config-data" (OuterVolumeSpecName: "config-data") pod "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" (UID: "f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.419923 4979 scope.go:117] "RemoveContainer" containerID="ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.454152 4979 scope.go:117] "RemoveContainer" containerID="968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.454666 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa\": container with ID starting with 968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa not found: ID does not exist" containerID="968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.454709 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa"} err="failed to get container status \"968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa\": rpc error: code = NotFound desc = could not find container \"968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa\": container with ID starting with 968ce15e83cec4a73162bdaf42b2e51fdecf8bcd5dda11f53c8abd2c4637dcaa not found: ID does not exist" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.454735 4979 scope.go:117] "RemoveContainer" containerID="51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.455292 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25\": container with ID starting with 51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25 not found: ID does not exist" containerID="51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.455348 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25"} err="failed to get container status \"51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25\": rpc error: code = NotFound desc = could not find container \"51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25\": container with ID starting with 51b1ab10783696504faeceb51fe179c934a5b3f1eaf3f1093862d8b6fbbe5a25 not found: ID does not exist" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.455378 4979 scope.go:117] "RemoveContainer" containerID="12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.455901 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612\": container with ID starting with 12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612 not found: ID does not exist" containerID="12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.455927 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612"} err="failed to get container status \"12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612\": rpc error: code = NotFound desc = could not find container \"12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612\": container with ID starting with 12d6155295e569ba4f19eee92d88a829d7838b408a7587522000024eca7d1612 not found: ID does not exist" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.455944 4979 scope.go:117] "RemoveContainer" containerID="ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.456190 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439\": container with ID starting with ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439 not found: ID does not exist" containerID="ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.456215 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439"} err="failed to get container status \"ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439\": rpc error: code = NotFound desc = could not find container \"ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439\": container with ID starting with ca7f8a4f77189ffa196c474255f354b197a5f83d952f542f40f86d8241372439 not found: ID does not exist" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.493535 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.624837 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.634065 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.650676 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.651211 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-central-agent" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651239 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-central-agent" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.651252 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="sg-core" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651258 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="sg-core" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.651287 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-notification-agent" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651308 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-notification-agent" Dec 04 12:05:05 crc kubenswrapper[4979]: E1204 12:05:05.651318 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="proxy-httpd" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651323 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="proxy-httpd" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651490 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-central-agent" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651505 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="ceilometer-notification-agent" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651519 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="sg-core" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.651531 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" containerName="proxy-httpd" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.653240 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.656511 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.656710 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.666124 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695320 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-run-httpd\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695363 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695405 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695433 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-log-httpd\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695452 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-scripts\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695501 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc276\" (UniqueName: \"kubernetes.io/projected/c067c12b-c393-4f9e-835c-dc2cf68ca81d-kube-api-access-gc276\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.695523 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-config-data\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.796762 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-config-data\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797411 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-run-httpd\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797483 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797511 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-log-httpd\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797534 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-scripts\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797580 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc276\" (UniqueName: \"kubernetes.io/projected/c067c12b-c393-4f9e-835c-dc2cf68ca81d-kube-api-access-gc276\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797911 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-log-httpd\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.797910 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-run-httpd\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.801684 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-config-data\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.802139 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.802782 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.805519 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-scripts\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.817070 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc276\" (UniqueName: \"kubernetes.io/projected/c067c12b-c393-4f9e-835c-dc2cf68ca81d-kube-api-access-gc276\") pod \"ceilometer-0\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " pod="openstack/ceilometer-0" Dec 04 12:05:05 crc kubenswrapper[4979]: I1204 12:05:05.973563 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:06 crc kubenswrapper[4979]: I1204 12:05:06.211255 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed" path="/var/lib/kubelet/pods/f3fc2ad1-f914-4bb8-b6fb-f6934a01c0ed/volumes" Dec 04 12:05:06 crc kubenswrapper[4979]: I1204 12:05:06.296656 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2465j" event={"ID":"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6","Type":"ContainerDied","Data":"07755584661f943a7836f5d09c354a242f8dc548a3c61640b8eac7753c254ff9"} Dec 04 12:05:06 crc kubenswrapper[4979]: I1204 12:05:06.296589 4979 generic.go:334] "Generic (PLEG): container finished" podID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" containerID="07755584661f943a7836f5d09c354a242f8dc548a3c61640b8eac7753c254ff9" exitCode=0 Dec 04 12:05:06 crc kubenswrapper[4979]: I1204 12:05:06.444867 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.326657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerStarted","Data":"5b2fd5dacf6062dc744a09b6022701e33feeaac8b72ebbd9515e937bbcad4ab3"} Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.327037 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerStarted","Data":"90e8366ddde0a9ab6445542c991b2a3ef4252b1aec16faa23b7fc68c62caa329"} Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.669563 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2465j" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850377 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-etc-machine-id\") pod \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850449 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-combined-ca-bundle\") pod \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850476 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-scripts\") pod \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850516 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" (UID: "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850593 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-db-sync-config-data\") pod \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850670 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-config-data\") pod \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.850697 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qp8l4\" (UniqueName: \"kubernetes.io/projected/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-kube-api-access-qp8l4\") pod \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\" (UID: \"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6\") " Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.851751 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.855128 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-kube-api-access-qp8l4" (OuterVolumeSpecName: "kube-api-access-qp8l4") pod "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" (UID: "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6"). InnerVolumeSpecName "kube-api-access-qp8l4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.855217 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" (UID: "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.857066 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-scripts" (OuterVolumeSpecName: "scripts") pod "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" (UID: "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.871866 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.877407 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" (UID: "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.904267 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-config-data" (OuterVolumeSpecName: "config-data") pod "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" (UID: "c0fd1874-37a5-40ab-ab1e-c18a0cffbab6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.952997 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.953156 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.953168 4979 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.953176 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:07 crc kubenswrapper[4979]: I1204 12:05:07.953184 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qp8l4\" (UniqueName: \"kubernetes.io/projected/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6-kube-api-access-qp8l4\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.337207 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerStarted","Data":"aae015188c0f6c9489c13750c36bce17c085f69fa8a8fcc1115a9eedc23611be"} Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.337251 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerStarted","Data":"19effdd5a13ee325f7551fc2ef7fb4ea8271ead6f804679e9588ad619f06fa37"} Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.340658 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-2465j" event={"ID":"c0fd1874-37a5-40ab-ab1e-c18a0cffbab6","Type":"ContainerDied","Data":"5c46d7ea59c4d94ad039c40d3eec69ed33119159d134390ee76c47be56200ea7"} Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.340853 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c46d7ea59c4d94ad039c40d3eec69ed33119159d134390ee76c47be56200ea7" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.341044 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-2465j" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.646846 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:08 crc kubenswrapper[4979]: E1204 12:05:08.647643 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" containerName="cinder-db-sync" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.647663 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" containerName="cinder-db-sync" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.647855 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" containerName="cinder-db-sync" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.648985 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.651056 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.652043 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.652324 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.652484 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-zf9zs" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.662968 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.677847 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b871793-8532-4ae9-998d-09d99a56f23c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.678101 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.678416 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.678509 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.678553 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.678574 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84q8b\" (UniqueName: \"kubernetes.io/projected/4b871793-8532-4ae9-998d-09d99a56f23c-kube-api-access-84q8b\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.771157 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qt9j9"] Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.771458 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerName="dnsmasq-dns" containerID="cri-o://b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86" gracePeriod=10 Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.777700 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.782155 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.782212 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.782237 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.782256 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84q8b\" (UniqueName: \"kubernetes.io/projected/4b871793-8532-4ae9-998d-09d99a56f23c-kube-api-access-84q8b\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.782363 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b871793-8532-4ae9-998d-09d99a56f23c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.783082 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b871793-8532-4ae9-998d-09d99a56f23c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.783454 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.791120 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-95sg6"] Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.793034 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.793220 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.793699 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.793740 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-scripts\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.802684 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.819675 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-95sg6"] Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.827401 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84q8b\" (UniqueName: \"kubernetes.io/projected/4b871793-8532-4ae9-998d-09d99a56f23c-kube-api-access-84q8b\") pod \"cinder-scheduler-0\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.886145 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.886330 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48plj\" (UniqueName: \"kubernetes.io/projected/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-kube-api-access-48plj\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.886574 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.886657 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.886714 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-config\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.886818 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.976962 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.988322 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.988584 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.988606 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-config\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.988636 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.988660 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.988698 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48plj\" (UniqueName: \"kubernetes.io/projected/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-kube-api-access-48plj\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.989844 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.990360 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.990883 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-config\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.991404 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.991883 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.997047 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:08 crc kubenswrapper[4979]: I1204 12:05:08.998521 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.006766 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.011595 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.047963 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48plj\" (UniqueName: \"kubernetes.io/projected/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-kube-api-access-48plj\") pod \"dnsmasq-dns-5c9776ccc5-95sg6\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091247 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-scripts\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091290 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56f16141-154b-401c-bf4e-ca6d7b73c040-etc-machine-id\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091325 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091382 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091429 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f16141-154b-401c-bf4e-ca6d7b73c040-logs\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091549 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data-custom\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.091593 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9prb\" (UniqueName: \"kubernetes.io/projected/56f16141-154b-401c-bf4e-ca6d7b73c040-kube-api-access-b9prb\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192536 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data-custom\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192578 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9prb\" (UniqueName: \"kubernetes.io/projected/56f16141-154b-401c-bf4e-ca6d7b73c040-kube-api-access-b9prb\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192594 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-scripts\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192613 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56f16141-154b-401c-bf4e-ca6d7b73c040-etc-machine-id\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192626 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192657 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.192684 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f16141-154b-401c-bf4e-ca6d7b73c040-logs\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.193148 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f16141-154b-401c-bf4e-ca6d7b73c040-logs\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.199560 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56f16141-154b-401c-bf4e-ca6d7b73c040-etc-machine-id\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.201037 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.201828 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-scripts\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.202328 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.205091 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data-custom\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.228427 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9prb\" (UniqueName: \"kubernetes.io/projected/56f16141-154b-401c-bf4e-ca6d7b73c040-kube-api-access-b9prb\") pod \"cinder-api-0\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.234077 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.306606 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.357922 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.365744 4979 generic.go:334] "Generic (PLEG): container finished" podID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerID="b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86" exitCode=0 Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.365782 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" event={"ID":"0c55939b-bdfd-4deb-8d49-11036a75f364","Type":"ContainerDied","Data":"b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86"} Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.365808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" event={"ID":"0c55939b-bdfd-4deb-8d49-11036a75f364","Type":"ContainerDied","Data":"3a3bf5fbfe356969e1159ee081d25588b6129f114e251015092d664cc43f7b8d"} Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.365824 4979 scope.go:117] "RemoveContainer" containerID="b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.365942 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-qt9j9" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.421532 4979 scope.go:117] "RemoveContainer" containerID="75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.447986 4979 scope.go:117] "RemoveContainer" containerID="b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86" Dec 04 12:05:09 crc kubenswrapper[4979]: E1204 12:05:09.450558 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86\": container with ID starting with b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86 not found: ID does not exist" containerID="b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.450591 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86"} err="failed to get container status \"b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86\": rpc error: code = NotFound desc = could not find container \"b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86\": container with ID starting with b6b4199eca3136350843077688905c3d009aeaa4750b1941feaa7f3fdbef9e86 not found: ID does not exist" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.450610 4979 scope.go:117] "RemoveContainer" containerID="75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c" Dec 04 12:05:09 crc kubenswrapper[4979]: E1204 12:05:09.451580 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c\": container with ID starting with 75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c not found: ID does not exist" containerID="75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.451628 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c"} err="failed to get container status \"75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c\": rpc error: code = NotFound desc = could not find container \"75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c\": container with ID starting with 75105b66852aadaa748eefcc1adcdbbad873c8e9a8ddaac9929896a043fe783c not found: ID does not exist" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.498888 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-nb\") pod \"0c55939b-bdfd-4deb-8d49-11036a75f364\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.498981 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-sb\") pod \"0c55939b-bdfd-4deb-8d49-11036a75f364\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.499007 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-svc\") pod \"0c55939b-bdfd-4deb-8d49-11036a75f364\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.499107 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpz8p\" (UniqueName: \"kubernetes.io/projected/0c55939b-bdfd-4deb-8d49-11036a75f364-kube-api-access-cpz8p\") pod \"0c55939b-bdfd-4deb-8d49-11036a75f364\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.499143 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-config\") pod \"0c55939b-bdfd-4deb-8d49-11036a75f364\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.499216 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-swift-storage-0\") pod \"0c55939b-bdfd-4deb-8d49-11036a75f364\" (UID: \"0c55939b-bdfd-4deb-8d49-11036a75f364\") " Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.520264 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c55939b-bdfd-4deb-8d49-11036a75f364-kube-api-access-cpz8p" (OuterVolumeSpecName: "kube-api-access-cpz8p") pod "0c55939b-bdfd-4deb-8d49-11036a75f364" (UID: "0c55939b-bdfd-4deb-8d49-11036a75f364"). InnerVolumeSpecName "kube-api-access-cpz8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.601231 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0c55939b-bdfd-4deb-8d49-11036a75f364" (UID: "0c55939b-bdfd-4deb-8d49-11036a75f364"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.601728 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpz8p\" (UniqueName: \"kubernetes.io/projected/0c55939b-bdfd-4deb-8d49-11036a75f364-kube-api-access-cpz8p\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.601748 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.605959 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.606852 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0c55939b-bdfd-4deb-8d49-11036a75f364" (UID: "0c55939b-bdfd-4deb-8d49-11036a75f364"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.622943 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-config" (OuterVolumeSpecName: "config") pod "0c55939b-bdfd-4deb-8d49-11036a75f364" (UID: "0c55939b-bdfd-4deb-8d49-11036a75f364"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:09 crc kubenswrapper[4979]: W1204 12:05:09.629771 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56f16141_154b_401c_bf4e_ca6d7b73c040.slice/crio-0b7750a43f82c0ac649580f6f5b7980743870a37d118f3fd1711b6d068d0d0ef WatchSource:0}: Error finding container 0b7750a43f82c0ac649580f6f5b7980743870a37d118f3fd1711b6d068d0d0ef: Status 404 returned error can't find the container with id 0b7750a43f82c0ac649580f6f5b7980743870a37d118f3fd1711b6d068d0d0ef Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.636436 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0c55939b-bdfd-4deb-8d49-11036a75f364" (UID: "0c55939b-bdfd-4deb-8d49-11036a75f364"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.658398 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.660309 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0c55939b-bdfd-4deb-8d49-11036a75f364" (UID: "0c55939b-bdfd-4deb-8d49-11036a75f364"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.705267 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.705362 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.705435 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.705450 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0c55939b-bdfd-4deb-8d49-11036a75f364-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.748473 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qt9j9"] Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.769062 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-qt9j9"] Dec 04 12:05:09 crc kubenswrapper[4979]: I1204 12:05:09.891907 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-95sg6"] Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.173597 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.228394 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" path="/var/lib/kubelet/pods/0c55939b-bdfd-4deb-8d49-11036a75f364/volumes" Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.233785 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58f74675fd-6lkn6"] Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.234000 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58f74675fd-6lkn6" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api-log" containerID="cri-o://a69924e3d2ab794ab8931fce1d8a12a718cbc7461f1b93f1393b6e1bcfd0c441" gracePeriod=30 Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.234415 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58f74675fd-6lkn6" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api" containerID="cri-o://11b2fee3fbc24534e5dea3a5dc9064a3b8e00d8dec1712ba02e67ce3f008705b" gracePeriod=30 Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.401691 4979 generic.go:334] "Generic (PLEG): container finished" podID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerID="2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a" exitCode=0 Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.402147 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" event={"ID":"b4c7c37c-3ab7-43db-ad20-8cd6df70395a","Type":"ContainerDied","Data":"2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a"} Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.402193 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" event={"ID":"b4c7c37c-3ab7-43db-ad20-8cd6df70395a","Type":"ContainerStarted","Data":"80f5dd7780f6820011c49d7d051b71260ba35d77206cc9882d5c9c7961c8a0fd"} Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.408350 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b871793-8532-4ae9-998d-09d99a56f23c","Type":"ContainerStarted","Data":"64879aa3e2ee782c1b68e5f48916db8c1d0589a0e06cba723a6465d45eb10eae"} Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.411477 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerStarted","Data":"0149a99d17e15b18db631076a86c39ff08e052d3f673de9b2be3392efd8aacd3"} Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.412356 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.414254 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"56f16141-154b-401c-bf4e-ca6d7b73c040","Type":"ContainerStarted","Data":"0b7750a43f82c0ac649580f6f5b7980743870a37d118f3fd1711b6d068d0d0ef"} Dec 04 12:05:10 crc kubenswrapper[4979]: I1204 12:05:10.456475 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.4743474340000002 podStartE2EDuration="5.456456806s" podCreationTimestamp="2025-12-04 12:05:05 +0000 UTC" firstStartedPulling="2025-12-04 12:05:06.455349408 +0000 UTC m=+1330.729645212" lastFinishedPulling="2025-12-04 12:05:09.43745878 +0000 UTC m=+1333.711754584" observedRunningTime="2025-12-04 12:05:10.446409673 +0000 UTC m=+1334.720705477" watchObservedRunningTime="2025-12-04 12:05:10.456456806 +0000 UTC m=+1334.730752630" Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.316666 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.472566 4979 generic.go:334] "Generic (PLEG): container finished" podID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerID="a69924e3d2ab794ab8931fce1d8a12a718cbc7461f1b93f1393b6e1bcfd0c441" exitCode=143 Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.472631 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58f74675fd-6lkn6" event={"ID":"dee1bb55-edbe-4cc0-989d-918bbd9dd487","Type":"ContainerDied","Data":"a69924e3d2ab794ab8931fce1d8a12a718cbc7461f1b93f1393b6e1bcfd0c441"} Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.498142 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"56f16141-154b-401c-bf4e-ca6d7b73c040","Type":"ContainerStarted","Data":"108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb"} Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.498187 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"56f16141-154b-401c-bf4e-ca6d7b73c040","Type":"ContainerStarted","Data":"4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de"} Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.498343 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api-log" containerID="cri-o://4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de" gracePeriod=30 Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.498586 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.498860 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api" containerID="cri-o://108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb" gracePeriod=30 Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.505392 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" event={"ID":"b4c7c37c-3ab7-43db-ad20-8cd6df70395a","Type":"ContainerStarted","Data":"26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6"} Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.505430 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.582122 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.5820924610000002 podStartE2EDuration="3.582092461s" podCreationTimestamp="2025-12-04 12:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:11.54454912 +0000 UTC m=+1335.818844924" watchObservedRunningTime="2025-12-04 12:05:11.582092461 +0000 UTC m=+1335.856388265" Dec 04 12:05:11 crc kubenswrapper[4979]: I1204 12:05:11.583340 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" podStartSLOduration=3.583331165 podStartE2EDuration="3.583331165s" podCreationTimestamp="2025-12-04 12:05:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:11.578950316 +0000 UTC m=+1335.853246130" watchObservedRunningTime="2025-12-04 12:05:11.583331165 +0000 UTC m=+1335.857626969" Dec 04 12:05:12 crc kubenswrapper[4979]: I1204 12:05:12.521657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b871793-8532-4ae9-998d-09d99a56f23c","Type":"ContainerStarted","Data":"b6e68d6ea0d1dd0c79750df49da474ebdac1c8d5f5803e3ed63315419354d383"} Dec 04 12:05:12 crc kubenswrapper[4979]: I1204 12:05:12.523920 4979 generic.go:334] "Generic (PLEG): container finished" podID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerID="4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de" exitCode=143 Dec 04 12:05:12 crc kubenswrapper[4979]: I1204 12:05:12.524005 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"56f16141-154b-401c-bf4e-ca6d7b73c040","Type":"ContainerDied","Data":"4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de"} Dec 04 12:05:13 crc kubenswrapper[4979]: I1204 12:05:13.565695 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b871793-8532-4ae9-998d-09d99a56f23c","Type":"ContainerStarted","Data":"a1159901ac037873bdb17caf79cd8c6474073f03e8897157d4bb27f4697d3086"} Dec 04 12:05:13 crc kubenswrapper[4979]: I1204 12:05:13.570168 4979 generic.go:334] "Generic (PLEG): container finished" podID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerID="11b2fee3fbc24534e5dea3a5dc9064a3b8e00d8dec1712ba02e67ce3f008705b" exitCode=0 Dec 04 12:05:13 crc kubenswrapper[4979]: I1204 12:05:13.570219 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58f74675fd-6lkn6" event={"ID":"dee1bb55-edbe-4cc0-989d-918bbd9dd487","Type":"ContainerDied","Data":"11b2fee3fbc24534e5dea3a5dc9064a3b8e00d8dec1712ba02e67ce3f008705b"} Dec 04 12:05:13 crc kubenswrapper[4979]: I1204 12:05:13.591640 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.40773341 podStartE2EDuration="5.591622669s" podCreationTimestamp="2025-12-04 12:05:08 +0000 UTC" firstStartedPulling="2025-12-04 12:05:09.668733078 +0000 UTC m=+1333.943028882" lastFinishedPulling="2025-12-04 12:05:10.852622337 +0000 UTC m=+1335.126918141" observedRunningTime="2025-12-04 12:05:13.587944289 +0000 UTC m=+1337.862240133" watchObservedRunningTime="2025-12-04 12:05:13.591622669 +0000 UTC m=+1337.865918473" Dec 04 12:05:13 crc kubenswrapper[4979]: I1204 12:05:13.954774 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:05:13 crc kubenswrapper[4979]: I1204 12:05:13.982471 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.003464 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dee1bb55-edbe-4cc0-989d-918bbd9dd487-logs\") pod \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.003529 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgnxh\" (UniqueName: \"kubernetes.io/projected/dee1bb55-edbe-4cc0-989d-918bbd9dd487-kube-api-access-dgnxh\") pod \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.003568 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-combined-ca-bundle\") pod \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.003654 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data\") pod \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.003681 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data-custom\") pod \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\" (UID: \"dee1bb55-edbe-4cc0-989d-918bbd9dd487\") " Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.004169 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dee1bb55-edbe-4cc0-989d-918bbd9dd487-logs" (OuterVolumeSpecName: "logs") pod "dee1bb55-edbe-4cc0-989d-918bbd9dd487" (UID: "dee1bb55-edbe-4cc0-989d-918bbd9dd487"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.013932 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dee1bb55-edbe-4cc0-989d-918bbd9dd487-kube-api-access-dgnxh" (OuterVolumeSpecName: "kube-api-access-dgnxh") pod "dee1bb55-edbe-4cc0-989d-918bbd9dd487" (UID: "dee1bb55-edbe-4cc0-989d-918bbd9dd487"). InnerVolumeSpecName "kube-api-access-dgnxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.027001 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dee1bb55-edbe-4cc0-989d-918bbd9dd487" (UID: "dee1bb55-edbe-4cc0-989d-918bbd9dd487"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.032022 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dee1bb55-edbe-4cc0-989d-918bbd9dd487" (UID: "dee1bb55-edbe-4cc0-989d-918bbd9dd487"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.051514 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data" (OuterVolumeSpecName: "config-data") pod "dee1bb55-edbe-4cc0-989d-918bbd9dd487" (UID: "dee1bb55-edbe-4cc0-989d-918bbd9dd487"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.105668 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dee1bb55-edbe-4cc0-989d-918bbd9dd487-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.105730 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgnxh\" (UniqueName: \"kubernetes.io/projected/dee1bb55-edbe-4cc0-989d-918bbd9dd487-kube-api-access-dgnxh\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.105754 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.105771 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.105788 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dee1bb55-edbe-4cc0-989d-918bbd9dd487-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.582379 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58f74675fd-6lkn6" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.582438 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58f74675fd-6lkn6" event={"ID":"dee1bb55-edbe-4cc0-989d-918bbd9dd487","Type":"ContainerDied","Data":"cffdab1c4f49fdc4bf399e736e5dc7c1dfdd72058185891a8eb144e96d678108"} Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.582587 4979 scope.go:117] "RemoveContainer" containerID="11b2fee3fbc24534e5dea3a5dc9064a3b8e00d8dec1712ba02e67ce3f008705b" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.614565 4979 scope.go:117] "RemoveContainer" containerID="a69924e3d2ab794ab8931fce1d8a12a718cbc7461f1b93f1393b6e1bcfd0c441" Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.618447 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58f74675fd-6lkn6"] Dec 04 12:05:14 crc kubenswrapper[4979]: I1204 12:05:14.635909 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-58f74675fd-6lkn6"] Dec 04 12:05:16 crc kubenswrapper[4979]: I1204 12:05:16.218343 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" path="/var/lib/kubelet/pods/dee1bb55-edbe-4cc0-989d-918bbd9dd487/volumes" Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.244229 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.308487 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.311174 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.323361 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.385228 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-95q7j"] Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.385473 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerName="dnsmasq-dns" containerID="cri-o://b47af92c762c0b6617f5114d10bd61c6e56f191d7fa672d3c0db1f97f616a2b8" gracePeriod=10 Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.649462 4979 generic.go:334] "Generic (PLEG): container finished" podID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerID="b47af92c762c0b6617f5114d10bd61c6e56f191d7fa672d3c0db1f97f616a2b8" exitCode=0 Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.649675 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" event={"ID":"edba8570-7f7d-4c8e-8917-862f5dcc59ba","Type":"ContainerDied","Data":"b47af92c762c0b6617f5114d10bd61c6e56f191d7fa672d3c0db1f97f616a2b8"} Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.649746 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="cinder-scheduler" containerID="cri-o://b6e68d6ea0d1dd0c79750df49da474ebdac1c8d5f5803e3ed63315419354d383" gracePeriod=30 Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.649799 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="probe" containerID="cri-o://a1159901ac037873bdb17caf79cd8c6474073f03e8897157d4bb27f4697d3086" gracePeriod=30 Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.945179 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:05:19 crc kubenswrapper[4979]: I1204 12:05:19.995197 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.020663 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-svc\") pod \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.020874 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-nb\") pod \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.020908 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-sb\") pod \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.020993 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-swift-storage-0\") pod \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.021050 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wzg4\" (UniqueName: \"kubernetes.io/projected/edba8570-7f7d-4c8e-8917-862f5dcc59ba-kube-api-access-9wzg4\") pod \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.021358 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-config\") pod \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\" (UID: \"edba8570-7f7d-4c8e-8917-862f5dcc59ba\") " Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.035696 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edba8570-7f7d-4c8e-8917-862f5dcc59ba-kube-api-access-9wzg4" (OuterVolumeSpecName: "kube-api-access-9wzg4") pod "edba8570-7f7d-4c8e-8917-862f5dcc59ba" (UID: "edba8570-7f7d-4c8e-8917-862f5dcc59ba"). InnerVolumeSpecName "kube-api-access-9wzg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.111854 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "edba8570-7f7d-4c8e-8917-862f5dcc59ba" (UID: "edba8570-7f7d-4c8e-8917-862f5dcc59ba"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.117979 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "edba8570-7f7d-4c8e-8917-862f5dcc59ba" (UID: "edba8570-7f7d-4c8e-8917-862f5dcc59ba"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.126633 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.126682 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.126695 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wzg4\" (UniqueName: \"kubernetes.io/projected/edba8570-7f7d-4c8e-8917-862f5dcc59ba-kube-api-access-9wzg4\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.137023 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "edba8570-7f7d-4c8e-8917-862f5dcc59ba" (UID: "edba8570-7f7d-4c8e-8917-862f5dcc59ba"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.154245 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-config" (OuterVolumeSpecName: "config") pod "edba8570-7f7d-4c8e-8917-862f5dcc59ba" (UID: "edba8570-7f7d-4c8e-8917-862f5dcc59ba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.167588 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "edba8570-7f7d-4c8e-8917-862f5dcc59ba" (UID: "edba8570-7f7d-4c8e-8917-862f5dcc59ba"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.228838 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.228881 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.228904 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/edba8570-7f7d-4c8e-8917-862f5dcc59ba-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.292847 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.457832 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.664491 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" event={"ID":"edba8570-7f7d-4c8e-8917-862f5dcc59ba","Type":"ContainerDied","Data":"0432b663e4218badb689c2c4381abbe2e497cb7c44008ca8322c11ab2a547f26"} Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.664839 4979 scope.go:117] "RemoveContainer" containerID="b47af92c762c0b6617f5114d10bd61c6e56f191d7fa672d3c0db1f97f616a2b8" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.664508 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-95q7j" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.668832 4979 generic.go:334] "Generic (PLEG): container finished" podID="4b871793-8532-4ae9-998d-09d99a56f23c" containerID="a1159901ac037873bdb17caf79cd8c6474073f03e8897157d4bb27f4697d3086" exitCode=0 Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.668865 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b871793-8532-4ae9-998d-09d99a56f23c","Type":"ContainerDied","Data":"a1159901ac037873bdb17caf79cd8c6474073f03e8897157d4bb27f4697d3086"} Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.688769 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-95q7j"] Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.692694 4979 scope.go:117] "RemoveContainer" containerID="8145e99f3ce77a36bd15c0b98448ad4bcb34344db4310d820df92958b5118f21" Dec 04 12:05:20 crc kubenswrapper[4979]: I1204 12:05:20.701867 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-95q7j"] Dec 04 12:05:21 crc kubenswrapper[4979]: I1204 12:05:21.436005 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 12:05:22 crc kubenswrapper[4979]: I1204 12:05:22.211150 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" path="/var/lib/kubelet/pods/edba8570-7f7d-4c8e-8917-862f5dcc59ba/volumes" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.714014 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 12:05:23 crc kubenswrapper[4979]: E1204 12:05:23.714864 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerName="dnsmasq-dns" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.714882 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerName="dnsmasq-dns" Dec 04 12:05:23 crc kubenswrapper[4979]: E1204 12:05:23.714904 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.714912 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api" Dec 04 12:05:23 crc kubenswrapper[4979]: E1204 12:05:23.714938 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api-log" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.714947 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api-log" Dec 04 12:05:23 crc kubenswrapper[4979]: E1204 12:05:23.714968 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerName="dnsmasq-dns" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.714975 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerName="dnsmasq-dns" Dec 04 12:05:23 crc kubenswrapper[4979]: E1204 12:05:23.714994 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerName="init" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.715001 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerName="init" Dec 04 12:05:23 crc kubenswrapper[4979]: E1204 12:05:23.715010 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerName="init" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.715017 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerName="init" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.715244 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api-log" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.715277 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="edba8570-7f7d-4c8e-8917-862f5dcc59ba" containerName="dnsmasq-dns" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.715292 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c55939b-bdfd-4deb-8d49-11036a75f364" containerName="dnsmasq-dns" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.715325 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dee1bb55-edbe-4cc0-989d-918bbd9dd487" containerName="barbican-api" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.716033 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.718593 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.718738 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.719216 4979 generic.go:334] "Generic (PLEG): container finished" podID="4b871793-8532-4ae9-998d-09d99a56f23c" containerID="b6e68d6ea0d1dd0c79750df49da474ebdac1c8d5f5803e3ed63315419354d383" exitCode=0 Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.719274 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b871793-8532-4ae9-998d-09d99a56f23c","Type":"ContainerDied","Data":"b6e68d6ea0d1dd0c79750df49da474ebdac1c8d5f5803e3ed63315419354d383"} Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.719788 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-jtw28" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.724960 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.904406 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dn9rq\" (UniqueName: \"kubernetes.io/projected/e8ecd6fb-35b6-4b0d-b236-71390feeed34-kube-api-access-dn9rq\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.904468 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.904543 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.904586 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config-secret\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:23 crc kubenswrapper[4979]: I1204 12:05:23.941139 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.006542 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config-secret\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.006622 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dn9rq\" (UniqueName: \"kubernetes.io/projected/e8ecd6fb-35b6-4b0d-b236-71390feeed34-kube-api-access-dn9rq\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.006665 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.006732 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.007655 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.012523 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-combined-ca-bundle\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.021890 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config-secret\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.024647 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dn9rq\" (UniqueName: \"kubernetes.io/projected/e8ecd6fb-35b6-4b0d-b236-71390feeed34-kube-api-access-dn9rq\") pod \"openstackclient\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.040891 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.107567 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data\") pod \"4b871793-8532-4ae9-998d-09d99a56f23c\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.108935 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84q8b\" (UniqueName: \"kubernetes.io/projected/4b871793-8532-4ae9-998d-09d99a56f23c-kube-api-access-84q8b\") pod \"4b871793-8532-4ae9-998d-09d99a56f23c\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.109005 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b871793-8532-4ae9-998d-09d99a56f23c-etc-machine-id\") pod \"4b871793-8532-4ae9-998d-09d99a56f23c\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.109026 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data-custom\") pod \"4b871793-8532-4ae9-998d-09d99a56f23c\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.109079 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b871793-8532-4ae9-998d-09d99a56f23c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4b871793-8532-4ae9-998d-09d99a56f23c" (UID: "4b871793-8532-4ae9-998d-09d99a56f23c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.109120 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-combined-ca-bundle\") pod \"4b871793-8532-4ae9-998d-09d99a56f23c\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.109757 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-scripts\") pod \"4b871793-8532-4ae9-998d-09d99a56f23c\" (UID: \"4b871793-8532-4ae9-998d-09d99a56f23c\") " Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.110473 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4b871793-8532-4ae9-998d-09d99a56f23c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.113669 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b871793-8532-4ae9-998d-09d99a56f23c-kube-api-access-84q8b" (OuterVolumeSpecName: "kube-api-access-84q8b") pod "4b871793-8532-4ae9-998d-09d99a56f23c" (UID: "4b871793-8532-4ae9-998d-09d99a56f23c"). InnerVolumeSpecName "kube-api-access-84q8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.113981 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4b871793-8532-4ae9-998d-09d99a56f23c" (UID: "4b871793-8532-4ae9-998d-09d99a56f23c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.114053 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-scripts" (OuterVolumeSpecName: "scripts") pod "4b871793-8532-4ae9-998d-09d99a56f23c" (UID: "4b871793-8532-4ae9-998d-09d99a56f23c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.199534 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b871793-8532-4ae9-998d-09d99a56f23c" (UID: "4b871793-8532-4ae9-998d-09d99a56f23c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.212213 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.212244 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84q8b\" (UniqueName: \"kubernetes.io/projected/4b871793-8532-4ae9-998d-09d99a56f23c-kube-api-access-84q8b\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.212257 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.212267 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.240816 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data" (OuterVolumeSpecName: "config-data") pod "4b871793-8532-4ae9-998d-09d99a56f23c" (UID: "4b871793-8532-4ae9-998d-09d99a56f23c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.313202 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b871793-8532-4ae9-998d-09d99a56f23c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.477323 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.477874 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.548793 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5cfc776c66-nvfsf"] Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.549238 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5cfc776c66-nvfsf" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-api" containerID="cri-o://2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b" gracePeriod=30 Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.549426 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5cfc776c66-nvfsf" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-httpd" containerID="cri-o://384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26" gracePeriod=30 Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.730601 4979 generic.go:334] "Generic (PLEG): container finished" podID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerID="384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26" exitCode=0 Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.730686 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cfc776c66-nvfsf" event={"ID":"30547b8b-d0c9-4473-9f34-f8b9bca3e20a","Type":"ContainerDied","Data":"384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26"} Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.733368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4b871793-8532-4ae9-998d-09d99a56f23c","Type":"ContainerDied","Data":"64879aa3e2ee782c1b68e5f48916db8c1d0589a0e06cba723a6465d45eb10eae"} Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.733426 4979 scope.go:117] "RemoveContainer" containerID="a1159901ac037873bdb17caf79cd8c6474073f03e8897157d4bb27f4697d3086" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.733586 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.738643 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e8ecd6fb-35b6-4b0d-b236-71390feeed34","Type":"ContainerStarted","Data":"33f7d774fd775234e26cc9bcd796f06e42fa835d93d72a48b7e9564ead6dc32b"} Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.777375 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.778346 4979 scope.go:117] "RemoveContainer" containerID="b6e68d6ea0d1dd0c79750df49da474ebdac1c8d5f5803e3ed63315419354d383" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.795188 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.815478 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:24 crc kubenswrapper[4979]: E1204 12:05:24.816050 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="cinder-scheduler" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.816075 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="cinder-scheduler" Dec 04 12:05:24 crc kubenswrapper[4979]: E1204 12:05:24.816117 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="probe" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.816127 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="probe" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.816543 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="probe" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.816559 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" containerName="cinder-scheduler" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.817768 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.819457 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.819490 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.819535 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.819590 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.819626 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgxkb\" (UniqueName: \"kubernetes.io/projected/010e4389-4656-4460-9b4d-2a7a37660252-kube-api-access-bgxkb\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.819673 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/010e4389-4656-4460-9b4d-2a7a37660252-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.820337 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.827363 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.921444 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.921776 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgxkb\" (UniqueName: \"kubernetes.io/projected/010e4389-4656-4460-9b4d-2a7a37660252-kube-api-access-bgxkb\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.921835 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/010e4389-4656-4460-9b4d-2a7a37660252-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.921863 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.921889 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.921927 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.922225 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/010e4389-4656-4460-9b4d-2a7a37660252-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.927588 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.930834 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.933727 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.934175 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:24 crc kubenswrapper[4979]: I1204 12:05:24.937194 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgxkb\" (UniqueName: \"kubernetes.io/projected/010e4389-4656-4460-9b4d-2a7a37660252-kube-api-access-bgxkb\") pod \"cinder-scheduler-0\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " pod="openstack/cinder-scheduler-0" Dec 04 12:05:25 crc kubenswrapper[4979]: I1204 12:05:25.132211 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:05:25 crc kubenswrapper[4979]: I1204 12:05:25.663478 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:05:25 crc kubenswrapper[4979]: I1204 12:05:25.756732 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"010e4389-4656-4460-9b4d-2a7a37660252","Type":"ContainerStarted","Data":"a7d0f5cde648670e4314265bd9d2c17a2a777fd27c1c26186aae34f8afd33964"} Dec 04 12:05:26 crc kubenswrapper[4979]: I1204 12:05:26.210393 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b871793-8532-4ae9-998d-09d99a56f23c" path="/var/lib/kubelet/pods/4b871793-8532-4ae9-998d-09d99a56f23c/volumes" Dec 04 12:05:26 crc kubenswrapper[4979]: I1204 12:05:26.776350 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"010e4389-4656-4460-9b4d-2a7a37660252","Type":"ContainerStarted","Data":"66265992f836c6a9f8e585ba56a557ed7f384463e217621e5223701c105f7441"} Dec 04 12:05:27 crc kubenswrapper[4979]: I1204 12:05:27.799669 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"010e4389-4656-4460-9b4d-2a7a37660252","Type":"ContainerStarted","Data":"aa4d07bf279adaae12d8aad041308e01478fd7338e4e1c8dfe3e93d2a095aa65"} Dec 04 12:05:27 crc kubenswrapper[4979]: I1204 12:05:27.819157 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.819135226 podStartE2EDuration="3.819135226s" podCreationTimestamp="2025-12-04 12:05:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:27.818285612 +0000 UTC m=+1352.092581416" watchObservedRunningTime="2025-12-04 12:05:27.819135226 +0000 UTC m=+1352.093431030" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.041202 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.041261 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.041325 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.042282 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb52f2da72fff7d717ea5b0066df40b5dc5b046752472d6fe21e232348fdf595"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.042370 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://fb52f2da72fff7d717ea5b0066df40b5dc5b046752472d6fe21e232348fdf595" gracePeriod=600 Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.050366 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-db585f97f-d5777"] Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.051881 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.053673 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.054023 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.054075 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.086425 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-db585f97f-d5777"] Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177558 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-etc-swift\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177640 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgpkl\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-kube-api-access-lgpkl\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177721 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-public-tls-certs\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177744 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-run-httpd\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177764 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-log-httpd\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177786 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-config-data\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177897 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-combined-ca-bundle\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.177940 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-internal-tls-certs\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.279282 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-public-tls-certs\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.279927 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-run-httpd\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.279958 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-log-httpd\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.279989 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-config-data\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.280037 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-combined-ca-bundle\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.280075 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-internal-tls-certs\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.280140 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-etc-swift\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.280193 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgpkl\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-kube-api-access-lgpkl\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.283847 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-log-httpd\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.284193 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-run-httpd\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.287909 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-public-tls-certs\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.288274 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-config-data\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.294874 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-internal-tls-certs\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.297824 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-etc-swift\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.302216 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-combined-ca-bundle\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.328857 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgpkl\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-kube-api-access-lgpkl\") pod \"swift-proxy-db585f97f-d5777\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.375811 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.457568 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.589893 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-httpd-config\") pod \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.589957 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-config\") pod \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.589974 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-combined-ca-bundle\") pod \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.590140 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-ovndb-tls-certs\") pod \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.590189 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x784c\" (UniqueName: \"kubernetes.io/projected/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-kube-api-access-x784c\") pod \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\" (UID: \"30547b8b-d0c9-4473-9f34-f8b9bca3e20a\") " Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.604470 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "30547b8b-d0c9-4473-9f34-f8b9bca3e20a" (UID: "30547b8b-d0c9-4473-9f34-f8b9bca3e20a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.604605 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-kube-api-access-x784c" (OuterVolumeSpecName: "kube-api-access-x784c") pod "30547b8b-d0c9-4473-9f34-f8b9bca3e20a" (UID: "30547b8b-d0c9-4473-9f34-f8b9bca3e20a"). InnerVolumeSpecName "kube-api-access-x784c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.686358 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30547b8b-d0c9-4473-9f34-f8b9bca3e20a" (UID: "30547b8b-d0c9-4473-9f34-f8b9bca3e20a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.692599 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x784c\" (UniqueName: \"kubernetes.io/projected/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-kube-api-access-x784c\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.692632 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.692644 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.700764 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-config" (OuterVolumeSpecName: "config") pod "30547b8b-d0c9-4473-9f34-f8b9bca3e20a" (UID: "30547b8b-d0c9-4473-9f34-f8b9bca3e20a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.724619 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "30547b8b-d0c9-4473-9f34-f8b9bca3e20a" (UID: "30547b8b-d0c9-4473-9f34-f8b9bca3e20a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.793764 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.793807 4979 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/30547b8b-d0c9-4473-9f34-f8b9bca3e20a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.812942 4979 generic.go:334] "Generic (PLEG): container finished" podID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerID="2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b" exitCode=0 Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.813009 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cfc776c66-nvfsf" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.813041 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cfc776c66-nvfsf" event={"ID":"30547b8b-d0c9-4473-9f34-f8b9bca3e20a","Type":"ContainerDied","Data":"2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b"} Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.813129 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cfc776c66-nvfsf" event={"ID":"30547b8b-d0c9-4473-9f34-f8b9bca3e20a","Type":"ContainerDied","Data":"5990fa32fb947afa7d83789dd1952f5630ba7467ccb7dcc3034d77d271c46919"} Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.813152 4979 scope.go:117] "RemoveContainer" containerID="384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.816401 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="fb52f2da72fff7d717ea5b0066df40b5dc5b046752472d6fe21e232348fdf595" exitCode=0 Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.816822 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"fb52f2da72fff7d717ea5b0066df40b5dc5b046752472d6fe21e232348fdf595"} Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.816877 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf"} Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.854649 4979 scope.go:117] "RemoveContainer" containerID="2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.863332 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5cfc776c66-nvfsf"] Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.872720 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5cfc776c66-nvfsf"] Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.881016 4979 scope.go:117] "RemoveContainer" containerID="384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26" Dec 04 12:05:28 crc kubenswrapper[4979]: E1204 12:05:28.882013 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26\": container with ID starting with 384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26 not found: ID does not exist" containerID="384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.882053 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26"} err="failed to get container status \"384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26\": rpc error: code = NotFound desc = could not find container \"384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26\": container with ID starting with 384c56635d9ab35591d812a9437d59bebae4342c9d6ab0d8aa826e29b4589e26 not found: ID does not exist" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.882101 4979 scope.go:117] "RemoveContainer" containerID="2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b" Dec 04 12:05:28 crc kubenswrapper[4979]: E1204 12:05:28.882623 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b\": container with ID starting with 2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b not found: ID does not exist" containerID="2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.882652 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b"} err="failed to get container status \"2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b\": rpc error: code = NotFound desc = could not find container \"2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b\": container with ID starting with 2716ed2188522581ef5cffc5260ee8414700fee5a613c50be7c28da2254e216b not found: ID does not exist" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.882677 4979 scope.go:117] "RemoveContainer" containerID="bb5b5fb8f2704900b621034dde12cd64efd7f6ba3338a4dd8572ae9e51dbc1d8" Dec 04 12:05:28 crc kubenswrapper[4979]: I1204 12:05:28.989756 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-db585f97f-d5777"] Dec 04 12:05:29 crc kubenswrapper[4979]: W1204 12:05:29.004629 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee632468_ca6f_48b2_abab_9cd939aafc00.slice/crio-9a87d1ecab55532e46e04e3cfc19b09df18880b41260ad47483b6cbddc18770a WatchSource:0}: Error finding container 9a87d1ecab55532e46e04e3cfc19b09df18880b41260ad47483b6cbddc18770a: Status 404 returned error can't find the container with id 9a87d1ecab55532e46e04e3cfc19b09df18880b41260ad47483b6cbddc18770a Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.630675 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.631531 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-central-agent" containerID="cri-o://5b2fd5dacf6062dc744a09b6022701e33feeaac8b72ebbd9515e937bbcad4ab3" gracePeriod=30 Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.631627 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="sg-core" containerID="cri-o://aae015188c0f6c9489c13750c36bce17c085f69fa8a8fcc1115a9eedc23611be" gracePeriod=30 Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.631672 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-notification-agent" containerID="cri-o://19effdd5a13ee325f7551fc2ef7fb4ea8271ead6f804679e9588ad619f06fa37" gracePeriod=30 Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.631684 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="proxy-httpd" containerID="cri-o://0149a99d17e15b18db631076a86c39ff08e052d3f673de9b2be3392efd8aacd3" gracePeriod=30 Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.639167 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.827100 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-db585f97f-d5777" event={"ID":"ee632468-ca6f-48b2-abab-9cd939aafc00","Type":"ContainerStarted","Data":"0c6ee4d9251618d455bcbaeca27562a8a00ff00726d03c52d5bbc4be6e44f194"} Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.827144 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-db585f97f-d5777" event={"ID":"ee632468-ca6f-48b2-abab-9cd939aafc00","Type":"ContainerStarted","Data":"17d10021029714eaf81e19c13f5d0dfddee111a3c2ce627763576fdf61297d0c"} Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.827155 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-db585f97f-d5777" event={"ID":"ee632468-ca6f-48b2-abab-9cd939aafc00","Type":"ContainerStarted","Data":"9a87d1ecab55532e46e04e3cfc19b09df18880b41260ad47483b6cbddc18770a"} Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.827362 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.827586 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.838451 4979 generic.go:334] "Generic (PLEG): container finished" podID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerID="aae015188c0f6c9489c13750c36bce17c085f69fa8a8fcc1115a9eedc23611be" exitCode=2 Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.838549 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerDied","Data":"aae015188c0f6c9489c13750c36bce17c085f69fa8a8fcc1115a9eedc23611be"} Dec 04 12:05:29 crc kubenswrapper[4979]: I1204 12:05:29.854018 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-db585f97f-d5777" podStartSLOduration=1.853999062 podStartE2EDuration="1.853999062s" podCreationTimestamp="2025-12-04 12:05:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:29.847214047 +0000 UTC m=+1354.121509851" watchObservedRunningTime="2025-12-04 12:05:29.853999062 +0000 UTC m=+1354.128294866" Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.133036 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.219389 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" path="/var/lib/kubelet/pods/30547b8b-d0c9-4473-9f34-f8b9bca3e20a/volumes" Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.251825 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.252117 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-log" containerID="cri-o://df98ae1983f5afe4ff2b92c1de05d7850f1596f0e50d699c019542f0b32b1c05" gracePeriod=30 Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.252187 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-httpd" containerID="cri-o://adf32560344263d5e4a29f6de9e9422adb452415702e0da15f1576b2842f89f9" gracePeriod=30 Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.868707 4979 generic.go:334] "Generic (PLEG): container finished" podID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerID="df98ae1983f5afe4ff2b92c1de05d7850f1596f0e50d699c019542f0b32b1c05" exitCode=143 Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.868808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"95a0e5e3-658f-4d68-9387-34966c2a8e1c","Type":"ContainerDied","Data":"df98ae1983f5afe4ff2b92c1de05d7850f1596f0e50d699c019542f0b32b1c05"} Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.875503 4979 generic.go:334] "Generic (PLEG): container finished" podID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerID="0149a99d17e15b18db631076a86c39ff08e052d3f673de9b2be3392efd8aacd3" exitCode=0 Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.875535 4979 generic.go:334] "Generic (PLEG): container finished" podID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerID="5b2fd5dacf6062dc744a09b6022701e33feeaac8b72ebbd9515e937bbcad4ab3" exitCode=0 Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.875569 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerDied","Data":"0149a99d17e15b18db631076a86c39ff08e052d3f673de9b2be3392efd8aacd3"} Dec 04 12:05:30 crc kubenswrapper[4979]: I1204 12:05:30.875607 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerDied","Data":"5b2fd5dacf6062dc744a09b6022701e33feeaac8b72ebbd9515e937bbcad4ab3"} Dec 04 12:05:33 crc kubenswrapper[4979]: I1204 12:05:33.413965 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.147:9292/healthcheck\": read tcp 10.217.0.2:45544->10.217.0.147:9292: read: connection reset by peer" Dec 04 12:05:33 crc kubenswrapper[4979]: I1204 12:05:33.413995 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.147:9292/healthcheck\": read tcp 10.217.0.2:45560->10.217.0.147:9292: read: connection reset by peer" Dec 04 12:05:33 crc kubenswrapper[4979]: I1204 12:05:33.908976 4979 generic.go:334] "Generic (PLEG): container finished" podID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerID="adf32560344263d5e4a29f6de9e9422adb452415702e0da15f1576b2842f89f9" exitCode=0 Dec 04 12:05:33 crc kubenswrapper[4979]: I1204 12:05:33.909291 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"95a0e5e3-658f-4d68-9387-34966c2a8e1c","Type":"ContainerDied","Data":"adf32560344263d5e4a29f6de9e9422adb452415702e0da15f1576b2842f89f9"} Dec 04 12:05:34 crc kubenswrapper[4979]: I1204 12:05:34.928341 4979 generic.go:334] "Generic (PLEG): container finished" podID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerID="19effdd5a13ee325f7551fc2ef7fb4ea8271ead6f804679e9588ad619f06fa37" exitCode=0 Dec 04 12:05:34 crc kubenswrapper[4979]: I1204 12:05:34.928657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerDied","Data":"19effdd5a13ee325f7551fc2ef7fb4ea8271ead6f804679e9588ad619f06fa37"} Dec 04 12:05:35 crc kubenswrapper[4979]: I1204 12:05:35.369314 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.039512 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.075853 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.878876761 podStartE2EDuration="13.075831909s" podCreationTimestamp="2025-12-04 12:05:23 +0000 UTC" firstStartedPulling="2025-12-04 12:05:24.481805345 +0000 UTC m=+1348.756101149" lastFinishedPulling="2025-12-04 12:05:35.678760493 +0000 UTC m=+1359.953056297" observedRunningTime="2025-12-04 12:05:35.977900757 +0000 UTC m=+1360.252196561" watchObservedRunningTime="2025-12-04 12:05:36.075831909 +0000 UTC m=+1360.350127713" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.112926 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.127656 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-config-data\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.127770 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-combined-ca-bundle\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.127872 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc276\" (UniqueName: \"kubernetes.io/projected/c067c12b-c393-4f9e-835c-dc2cf68ca81d-kube-api-access-gc276\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.127984 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-log-httpd\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.128010 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-scripts\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.128040 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-sg-core-conf-yaml\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.128081 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-run-httpd\") pod \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\" (UID: \"c067c12b-c393-4f9e-835c-dc2cf68ca81d\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.130120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.130390 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.135757 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.135816 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c067c12b-c393-4f9e-835c-dc2cf68ca81d-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.139424 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-scripts" (OuterVolumeSpecName: "scripts") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.165464 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c067c12b-c393-4f9e-835c-dc2cf68ca81d-kube-api-access-gc276" (OuterVolumeSpecName: "kube-api-access-gc276") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "kube-api-access-gc276". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.206377 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237026 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-httpd-run\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237094 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-scripts\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237145 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-logs\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237194 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-internal-tls-certs\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237268 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-config-data\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237325 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237393 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-combined-ca-bundle\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237469 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skg7k\" (UniqueName: \"kubernetes.io/projected/95a0e5e3-658f-4d68-9387-34966c2a8e1c-kube-api-access-skg7k\") pod \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\" (UID: \"95a0e5e3-658f-4d68-9387-34966c2a8e1c\") " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.237978 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.238002 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.238017 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc276\" (UniqueName: \"kubernetes.io/projected/c067c12b-c393-4f9e-835c-dc2cf68ca81d-kube-api-access-gc276\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.239689 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.241266 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-logs" (OuterVolumeSpecName: "logs") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.245026 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-scripts" (OuterVolumeSpecName: "scripts") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.245134 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.245404 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95a0e5e3-658f-4d68-9387-34966c2a8e1c-kube-api-access-skg7k" (OuterVolumeSpecName: "kube-api-access-skg7k") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "kube-api-access-skg7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.250562 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.272772 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-config-data" (OuterVolumeSpecName: "config-data") pod "c067c12b-c393-4f9e-835c-dc2cf68ca81d" (UID: "c067c12b-c393-4f9e-835c-dc2cf68ca81d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.274209 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.303145 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.306235 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-config-data" (OuterVolumeSpecName: "config-data") pod "95a0e5e3-658f-4d68-9387-34966c2a8e1c" (UID: "95a0e5e3-658f-4d68-9387-34966c2a8e1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339412 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339467 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339478 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339487 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339520 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339529 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c067c12b-c393-4f9e-835c-dc2cf68ca81d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339539 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339548 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skg7k\" (UniqueName: \"kubernetes.io/projected/95a0e5e3-658f-4d68-9387-34966c2a8e1c-kube-api-access-skg7k\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339558 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/95a0e5e3-658f-4d68-9387-34966c2a8e1c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.339565 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95a0e5e3-658f-4d68-9387-34966c2a8e1c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.361971 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.441153 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.795097 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.795554 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-log" containerID="cri-o://c29efb06e4e3649086453955662a6ccf8e8a5eb9f46956a25e0d23c4118a24f7" gracePeriod=30 Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.795676 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-httpd" containerID="cri-o://78cc8bc9044d8fcae9cd85ae7b724b226af0401ec3dd0144a560f7e93ebe8c61" gracePeriod=30 Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.966556 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c067c12b-c393-4f9e-835c-dc2cf68ca81d","Type":"ContainerDied","Data":"90e8366ddde0a9ab6445542c991b2a3ef4252b1aec16faa23b7fc68c62caa329"} Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.966614 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.966622 4979 scope.go:117] "RemoveContainer" containerID="0149a99d17e15b18db631076a86c39ff08e052d3f673de9b2be3392efd8aacd3" Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.971363 4979 generic.go:334] "Generic (PLEG): container finished" podID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerID="c29efb06e4e3649086453955662a6ccf8e8a5eb9f46956a25e0d23c4118a24f7" exitCode=143 Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.971440 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7","Type":"ContainerDied","Data":"c29efb06e4e3649086453955662a6ccf8e8a5eb9f46956a25e0d23c4118a24f7"} Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.973540 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e8ecd6fb-35b6-4b0d-b236-71390feeed34","Type":"ContainerStarted","Data":"7c53f35962685fc14e259872b8140e5617d6f353e0dc5a20a4f05710bcea99f1"} Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.981524 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"95a0e5e3-658f-4d68-9387-34966c2a8e1c","Type":"ContainerDied","Data":"bc4051c92354e8478fa59ca47517a8b38dab985f5003f9ef117ddcbab034a529"} Dec 04 12:05:36 crc kubenswrapper[4979]: I1204 12:05:36.981601 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.008587 4979 scope.go:117] "RemoveContainer" containerID="aae015188c0f6c9489c13750c36bce17c085f69fa8a8fcc1115a9eedc23611be" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.034452 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.040520 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.043080 4979 scope.go:117] "RemoveContainer" containerID="19effdd5a13ee325f7551fc2ef7fb4ea8271ead6f804679e9588ad619f06fa37" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.058492 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.075664 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091189 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091721 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="proxy-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091746 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="proxy-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091766 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-central-agent" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091775 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-central-agent" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091786 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091796 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091809 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-notification-agent" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091817 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-notification-agent" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091835 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-log" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091842 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-log" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091851 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="sg-core" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091858 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="sg-core" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091870 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091877 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: E1204 12:05:37.091925 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-api" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.091933 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-api" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092123 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-central-agent" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092143 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-log" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092160 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092166 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="ceilometer-notification-agent" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092184 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="proxy-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092195 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" containerName="glance-httpd" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092212 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="sg-core" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.092225 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="30547b8b-d0c9-4473-9f34-f8b9bca3e20a" containerName="neutron-api" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.094172 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.096230 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.096524 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.101227 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.107726 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.110139 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.112158 4979 scope.go:117] "RemoveContainer" containerID="5b2fd5dacf6062dc744a09b6022701e33feeaac8b72ebbd9515e937bbcad4ab3" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.112509 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.112785 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.115734 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154616 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154674 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154712 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154744 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-scripts\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154786 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8wjs\" (UniqueName: \"kubernetes.io/projected/bd84d754-4092-4b2c-a8aa-6727c88c5100-kube-api-access-d8wjs\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154814 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154830 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-config-data\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154866 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-logs\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154884 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154916 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154945 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6zcf\" (UniqueName: \"kubernetes.io/projected/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-kube-api-access-w6zcf\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154970 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.154985 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-log-httpd\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.155004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-run-httpd\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.155027 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.159858 4979 scope.go:117] "RemoveContainer" containerID="adf32560344263d5e4a29f6de9e9422adb452415702e0da15f1576b2842f89f9" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.184522 4979 scope.go:117] "RemoveContainer" containerID="df98ae1983f5afe4ff2b92c1de05d7850f1596f0e50d699c019542f0b32b1c05" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.256765 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.257367 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-log-httpd\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.257495 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-run-httpd\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.257625 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.257763 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.257921 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258065 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258630 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-scripts\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258784 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8wjs\" (UniqueName: \"kubernetes.io/projected/bd84d754-4092-4b2c-a8aa-6727c88c5100-kube-api-access-d8wjs\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258901 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258996 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-config-data\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.259155 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-logs\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.259267 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.259405 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.259542 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-log-httpd\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.257639 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.259544 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6zcf\" (UniqueName: \"kubernetes.io/projected/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-kube-api-access-w6zcf\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258359 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.258117 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-run-httpd\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.260536 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-logs\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.264053 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-scripts\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.264145 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.264320 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.265933 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.266059 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.271588 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-config-data\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.273402 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-config-data\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.275134 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-scripts\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.275902 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8wjs\" (UniqueName: \"kubernetes.io/projected/bd84d754-4092-4b2c-a8aa-6727c88c5100-kube-api-access-d8wjs\") pod \"ceilometer-0\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.292765 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6zcf\" (UniqueName: \"kubernetes.io/projected/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-kube-api-access-w6zcf\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.297856 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.425904 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.437573 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.904261 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: W1204 12:05:37.989735 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2d24f5a_ae53_44e3_bc4c_c56813f0f141.slice/crio-a2ff341b559a4aa3d9cc2c4f5310f92c3beb84d8e92027d88499f5318b3249a4 WatchSource:0}: Error finding container a2ff341b559a4aa3d9cc2c4f5310f92c3beb84d8e92027d88499f5318b3249a4: Status 404 returned error can't find the container with id a2ff341b559a4aa3d9cc2c4f5310f92c3beb84d8e92027d88499f5318b3249a4 Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.991030 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:05:37 crc kubenswrapper[4979]: I1204 12:05:37.995457 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerStarted","Data":"9630538490cfa2913bdaff726a116bbe0fe6ef8225fe35826e5a7dd7c0b6e129"} Dec 04 12:05:38 crc kubenswrapper[4979]: I1204 12:05:38.211143 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95a0e5e3-658f-4d68-9387-34966c2a8e1c" path="/var/lib/kubelet/pods/95a0e5e3-658f-4d68-9387-34966c2a8e1c/volumes" Dec 04 12:05:38 crc kubenswrapper[4979]: I1204 12:05:38.211971 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" path="/var/lib/kubelet/pods/c067c12b-c393-4f9e-835c-dc2cf68ca81d/volumes" Dec 04 12:05:38 crc kubenswrapper[4979]: I1204 12:05:38.389138 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:38 crc kubenswrapper[4979]: I1204 12:05:38.393788 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:05:39 crc kubenswrapper[4979]: I1204 12:05:39.036008 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerStarted","Data":"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608"} Dec 04 12:05:39 crc kubenswrapper[4979]: I1204 12:05:39.039346 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d24f5a-ae53-44e3-bc4c-c56813f0f141","Type":"ContainerStarted","Data":"0b18da01797ecd4c68f4b61b2e504256585316a4b410185ba40df8298d3b3107"} Dec 04 12:05:39 crc kubenswrapper[4979]: I1204 12:05:39.039378 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d24f5a-ae53-44e3-bc4c-c56813f0f141","Type":"ContainerStarted","Data":"a2ff341b559a4aa3d9cc2c4f5310f92c3beb84d8e92027d88499f5318b3249a4"} Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.059472 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerStarted","Data":"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc"} Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.059982 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerStarted","Data":"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893"} Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.063855 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d24f5a-ae53-44e3-bc4c-c56813f0f141","Type":"ContainerStarted","Data":"02024a485457048e2f54ce565223a36a682c9da052f1a487dff4004a22a382ee"} Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.073728 4979 generic.go:334] "Generic (PLEG): container finished" podID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerID="78cc8bc9044d8fcae9cd85ae7b724b226af0401ec3dd0144a560f7e93ebe8c61" exitCode=0 Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.073769 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7","Type":"ContainerDied","Data":"78cc8bc9044d8fcae9cd85ae7b724b226af0401ec3dd0144a560f7e93ebe8c61"} Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.087098 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.087080313 podStartE2EDuration="3.087080313s" podCreationTimestamp="2025-12-04 12:05:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:40.080704929 +0000 UTC m=+1364.355000733" watchObservedRunningTime="2025-12-04 12:05:40.087080313 +0000 UTC m=+1364.361376117" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.254755 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.401654 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.545707 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-scripts\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.545763 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-public-tls-certs\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.545891 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-logs\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.545929 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-config-data\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.546056 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-combined-ca-bundle\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.546099 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-httpd-run\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.546179 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.546228 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlf5t\" (UniqueName: \"kubernetes.io/projected/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-kube-api-access-qlf5t\") pod \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\" (UID: \"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7\") " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.547059 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.547774 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-logs" (OuterVolumeSpecName: "logs") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.552010 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-kube-api-access-qlf5t" (OuterVolumeSpecName: "kube-api-access-qlf5t") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "kube-api-access-qlf5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.552218 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-scripts" (OuterVolumeSpecName: "scripts") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.568539 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.593540 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.606409 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.608502 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-config-data" (OuterVolumeSpecName: "config-data") pod "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" (UID: "3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647879 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlf5t\" (UniqueName: \"kubernetes.io/projected/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-kube-api-access-qlf5t\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647917 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647928 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647936 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647944 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647953 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647962 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.647996 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.667701 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 12:05:40 crc kubenswrapper[4979]: I1204 12:05:40.749258 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.086476 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7","Type":"ContainerDied","Data":"54d11babcbe0046ee19be2a2ad71fb79be3d97a1b7f1d43e1a47dcc8b2c35871"} Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.087515 4979 scope.go:117] "RemoveContainer" containerID="78cc8bc9044d8fcae9cd85ae7b724b226af0401ec3dd0144a560f7e93ebe8c61" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.086560 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.117632 4979 scope.go:117] "RemoveContainer" containerID="c29efb06e4e3649086453955662a6ccf8e8a5eb9f46956a25e0d23c4118a24f7" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.119433 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.127719 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.151356 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:05:41 crc kubenswrapper[4979]: E1204 12:05:41.151771 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-log" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.151787 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-log" Dec 04 12:05:41 crc kubenswrapper[4979]: E1204 12:05:41.151795 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-httpd" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.151802 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-httpd" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.151992 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-httpd" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.152018 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" containerName="glance-log" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.152889 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.160104 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.160339 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.181990 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258370 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-config-data\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258446 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-logs\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258485 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-scripts\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258509 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258538 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258614 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258656 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccfcl\" (UniqueName: \"kubernetes.io/projected/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-kube-api-access-ccfcl\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.258684 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.360431 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-config-data\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361022 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-logs\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361136 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-scripts\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361209 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361280 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361436 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361587 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-logs\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccfcl\" (UniqueName: \"kubernetes.io/projected/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-kube-api-access-ccfcl\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361849 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361761 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.361762 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.365158 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.366255 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-config-data\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.368831 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.371823 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-scripts\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.380326 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccfcl\" (UniqueName: \"kubernetes.io/projected/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-kube-api-access-ccfcl\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.394686 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-external-api-0\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " pod="openstack/glance-default-external-api-0" Dec 04 12:05:41 crc kubenswrapper[4979]: I1204 12:05:41.478327 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.002781 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.067496 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.109684 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c","Type":"ContainerStarted","Data":"39910816ab33a8692857530e858fc019709d1d1558e496405b3d4da777c1dc83"} Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.120240 4979 generic.go:334] "Generic (PLEG): container finished" podID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerID="108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb" exitCode=137 Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.120291 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"56f16141-154b-401c-bf4e-ca6d7b73c040","Type":"ContainerDied","Data":"108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb"} Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.120328 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"56f16141-154b-401c-bf4e-ca6d7b73c040","Type":"ContainerDied","Data":"0b7750a43f82c0ac649580f6f5b7980743870a37d118f3fd1711b6d068d0d0ef"} Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.120346 4979 scope.go:117] "RemoveContainer" containerID="108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.120431 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.130604 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerStarted","Data":"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0"} Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.130756 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-central-agent" containerID="cri-o://d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" gracePeriod=30 Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.130804 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.130814 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="sg-core" containerID="cri-o://400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" gracePeriod=30 Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.130789 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="proxy-httpd" containerID="cri-o://72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" gracePeriod=30 Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.130870 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-notification-agent" containerID="cri-o://6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" gracePeriod=30 Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.153938 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.8530385090000001 podStartE2EDuration="5.153916848s" podCreationTimestamp="2025-12-04 12:05:37 +0000 UTC" firstStartedPulling="2025-12-04 12:05:37.915022075 +0000 UTC m=+1362.189317899" lastFinishedPulling="2025-12-04 12:05:41.215900444 +0000 UTC m=+1365.490196238" observedRunningTime="2025-12-04 12:05:42.149382505 +0000 UTC m=+1366.423678319" watchObservedRunningTime="2025-12-04 12:05:42.153916848 +0000 UTC m=+1366.428212652" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177002 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-combined-ca-bundle\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177407 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9prb\" (UniqueName: \"kubernetes.io/projected/56f16141-154b-401c-bf4e-ca6d7b73c040-kube-api-access-b9prb\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177460 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177492 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f16141-154b-401c-bf4e-ca6d7b73c040-logs\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177608 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-scripts\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177654 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data-custom\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.177732 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56f16141-154b-401c-bf4e-ca6d7b73c040-etc-machine-id\") pod \"56f16141-154b-401c-bf4e-ca6d7b73c040\" (UID: \"56f16141-154b-401c-bf4e-ca6d7b73c040\") " Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.178837 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/56f16141-154b-401c-bf4e-ca6d7b73c040-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.179624 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56f16141-154b-401c-bf4e-ca6d7b73c040-logs" (OuterVolumeSpecName: "logs") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.182276 4979 scope.go:117] "RemoveContainer" containerID="4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.184774 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-scripts" (OuterVolumeSpecName: "scripts") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.185147 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56f16141-154b-401c-bf4e-ca6d7b73c040-kube-api-access-b9prb" (OuterVolumeSpecName: "kube-api-access-b9prb") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "kube-api-access-b9prb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.190999 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.215328 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.224994 4979 scope.go:117] "RemoveContainer" containerID="108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb" Dec 04 12:05:42 crc kubenswrapper[4979]: E1204 12:05:42.229043 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb\": container with ID starting with 108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb not found: ID does not exist" containerID="108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.230137 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb"} err="failed to get container status \"108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb\": rpc error: code = NotFound desc = could not find container \"108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb\": container with ID starting with 108b1110ea5b535e0e6b61881e9c2aed5e7b2a4d4f82504beeabea0942a1fffb not found: ID does not exist" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.230186 4979 scope.go:117] "RemoveContainer" containerID="4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de" Dec 04 12:05:42 crc kubenswrapper[4979]: E1204 12:05:42.232480 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de\": container with ID starting with 4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de not found: ID does not exist" containerID="4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.236081 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de"} err="failed to get container status \"4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de\": rpc error: code = NotFound desc = could not find container \"4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de\": container with ID starting with 4dca526dd7c653c412e7e38297d5366e3efb17308b9084b4c6b517d4a18b26de not found: ID does not exist" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.269384 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7" path="/var/lib/kubelet/pods/3debaaed-ce3f-4bff-a2d5-2cbc9d6e3db7/volumes" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.280373 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.280589 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.280705 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/56f16141-154b-401c-bf4e-ca6d7b73c040-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.280775 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.280839 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9prb\" (UniqueName: \"kubernetes.io/projected/56f16141-154b-401c-bf4e-ca6d7b73c040-kube-api-access-b9prb\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.280917 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/56f16141-154b-401c-bf4e-ca6d7b73c040-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.289451 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data" (OuterVolumeSpecName: "config-data") pod "56f16141-154b-401c-bf4e-ca6d7b73c040" (UID: "56f16141-154b-401c-bf4e-ca6d7b73c040"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.392394 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/56f16141-154b-401c-bf4e-ca6d7b73c040-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.450437 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.460051 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.481109 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:42 crc kubenswrapper[4979]: E1204 12:05:42.481563 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.481585 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api" Dec 04 12:05:42 crc kubenswrapper[4979]: E1204 12:05:42.481638 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api-log" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.481645 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api-log" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.481818 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.481844 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" containerName="cinder-api-log" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.482834 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.484796 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.486171 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.486361 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.490507 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595329 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595393 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-scripts\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595464 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ad703ff-c521-40f1-9fa4-09fca7726079-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595502 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ad703ff-c521-40f1-9fa4-09fca7726079-logs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595519 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595561 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxptx\" (UniqueName: \"kubernetes.io/projected/4ad703ff-c521-40f1-9fa4-09fca7726079-kube-api-access-cxptx\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595579 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595609 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.595638 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data-custom\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.698994 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-scripts\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699387 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ad703ff-c521-40f1-9fa4-09fca7726079-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699435 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ad703ff-c521-40f1-9fa4-09fca7726079-logs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699454 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699490 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxptx\" (UniqueName: \"kubernetes.io/projected/4ad703ff-c521-40f1-9fa4-09fca7726079-kube-api-access-cxptx\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699506 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699530 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699555 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data-custom\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.699605 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.700547 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ad703ff-c521-40f1-9fa4-09fca7726079-logs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.700621 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ad703ff-c521-40f1-9fa4-09fca7726079-etc-machine-id\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.704606 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-scripts\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.705440 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.705915 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.705985 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data-custom\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.707814 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.708493 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-public-tls-certs\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.715800 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxptx\" (UniqueName: \"kubernetes.io/projected/4ad703ff-c521-40f1-9fa4-09fca7726079-kube-api-access-cxptx\") pod \"cinder-api-0\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " pod="openstack/cinder-api-0" Dec 04 12:05:42 crc kubenswrapper[4979]: I1204 12:05:42.848373 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.117355 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155063 4979 generic.go:334] "Generic (PLEG): container finished" podID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" exitCode=0 Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155117 4979 generic.go:334] "Generic (PLEG): container finished" podID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" exitCode=2 Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155125 4979 generic.go:334] "Generic (PLEG): container finished" podID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" exitCode=0 Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155132 4979 generic.go:334] "Generic (PLEG): container finished" podID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" exitCode=0 Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155256 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155241 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerDied","Data":"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0"} Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155349 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerDied","Data":"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc"} Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155364 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerDied","Data":"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893"} Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155377 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerDied","Data":"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608"} Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155411 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bd84d754-4092-4b2c-a8aa-6727c88c5100","Type":"ContainerDied","Data":"9630538490cfa2913bdaff726a116bbe0fe6ef8225fe35826e5a7dd7c0b6e129"} Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.155431 4979 scope.go:117] "RemoveContainer" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.157488 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c","Type":"ContainerStarted","Data":"7341c4702a0b8160bf9a2a2ae8f818d6b9089d0a90f6a4ed29d2510863a765d0"} Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.199785 4979 scope.go:117] "RemoveContainer" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210523 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-combined-ca-bundle\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210621 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-log-httpd\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210652 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8wjs\" (UniqueName: \"kubernetes.io/projected/bd84d754-4092-4b2c-a8aa-6727c88c5100-kube-api-access-d8wjs\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210681 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-sg-core-conf-yaml\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210786 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-scripts\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210820 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-run-httpd\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.210861 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-config-data\") pod \"bd84d754-4092-4b2c-a8aa-6727c88c5100\" (UID: \"bd84d754-4092-4b2c-a8aa-6727c88c5100\") " Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.211765 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.211892 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.215131 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-scripts" (OuterVolumeSpecName: "scripts") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.215462 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd84d754-4092-4b2c-a8aa-6727c88c5100-kube-api-access-d8wjs" (OuterVolumeSpecName: "kube-api-access-d8wjs") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "kube-api-access-d8wjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.238575 4979 scope.go:117] "RemoveContainer" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.251515 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.262132 4979 scope.go:117] "RemoveContainer" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.288985 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.294945 4979 scope.go:117] "RemoveContainer" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.295488 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": container with ID starting with 72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0 not found: ID does not exist" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.295554 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0"} err="failed to get container status \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": rpc error: code = NotFound desc = could not find container \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": container with ID starting with 72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.295586 4979 scope.go:117] "RemoveContainer" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.296026 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": container with ID starting with 400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc not found: ID does not exist" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.296072 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc"} err="failed to get container status \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": rpc error: code = NotFound desc = could not find container \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": container with ID starting with 400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.296128 4979 scope.go:117] "RemoveContainer" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.296671 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": container with ID starting with 6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893 not found: ID does not exist" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.296723 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893"} err="failed to get container status \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": rpc error: code = NotFound desc = could not find container \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": container with ID starting with 6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.296748 4979 scope.go:117] "RemoveContainer" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.297519 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": container with ID starting with d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608 not found: ID does not exist" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.297820 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608"} err="failed to get container status \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": rpc error: code = NotFound desc = could not find container \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": container with ID starting with d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.297861 4979 scope.go:117] "RemoveContainer" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.298177 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0"} err="failed to get container status \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": rpc error: code = NotFound desc = could not find container \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": container with ID starting with 72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.298207 4979 scope.go:117] "RemoveContainer" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.302636 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc"} err="failed to get container status \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": rpc error: code = NotFound desc = could not find container \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": container with ID starting with 400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.302688 4979 scope.go:117] "RemoveContainer" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.303131 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893"} err="failed to get container status \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": rpc error: code = NotFound desc = could not find container \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": container with ID starting with 6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.303167 4979 scope.go:117] "RemoveContainer" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.303592 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608"} err="failed to get container status \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": rpc error: code = NotFound desc = could not find container \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": container with ID starting with d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.303625 4979 scope.go:117] "RemoveContainer" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.306373 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0"} err="failed to get container status \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": rpc error: code = NotFound desc = could not find container \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": container with ID starting with 72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.306400 4979 scope.go:117] "RemoveContainer" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.306781 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc"} err="failed to get container status \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": rpc error: code = NotFound desc = could not find container \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": container with ID starting with 400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.306806 4979 scope.go:117] "RemoveContainer" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307118 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893"} err="failed to get container status \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": rpc error: code = NotFound desc = could not find container \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": container with ID starting with 6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307136 4979 scope.go:117] "RemoveContainer" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307371 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608"} err="failed to get container status \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": rpc error: code = NotFound desc = could not find container \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": container with ID starting with d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307398 4979 scope.go:117] "RemoveContainer" containerID="72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307601 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0"} err="failed to get container status \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": rpc error: code = NotFound desc = could not find container \"72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0\": container with ID starting with 72f82ea6bf816ade453e34df5d1c03e766f89b81c46487c497c869ea45d327c0 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307631 4979 scope.go:117] "RemoveContainer" containerID="400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307879 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc"} err="failed to get container status \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": rpc error: code = NotFound desc = could not find container \"400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc\": container with ID starting with 400bc90a2dff9d54edf61d9f3cb09abd053e3bf5253a98814244a8e3116b8cfc not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.307902 4979 scope.go:117] "RemoveContainer" containerID="6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.308193 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893"} err="failed to get container status \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": rpc error: code = NotFound desc = could not find container \"6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893\": container with ID starting with 6f29d63d7744a40178fdcf38245d1ab526125cfea6745f125e7415bf94aab893 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.308209 4979 scope.go:117] "RemoveContainer" containerID="d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.308556 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608"} err="failed to get container status \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": rpc error: code = NotFound desc = could not find container \"d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608\": container with ID starting with d030d3b2c21564458cf99431c1dc7d9c92b2683c36990375806600c18fe22608 not found: ID does not exist" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.312642 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.312665 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8wjs\" (UniqueName: \"kubernetes.io/projected/bd84d754-4092-4b2c-a8aa-6727c88c5100-kube-api-access-d8wjs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.312679 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.312691 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.312701 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bd84d754-4092-4b2c-a8aa-6727c88c5100-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.312713 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.324162 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-config-data" (OuterVolumeSpecName: "config-data") pod "bd84d754-4092-4b2c-a8aa-6727c88c5100" (UID: "bd84d754-4092-4b2c-a8aa-6727c88c5100"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.357271 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:05:43 crc kubenswrapper[4979]: W1204 12:05:43.367260 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ad703ff_c521_40f1_9fa4_09fca7726079.slice/crio-bafb6eed795747c7239e30c7c1fd37c82ec310d9273eca36e1013feae0910bb7 WatchSource:0}: Error finding container bafb6eed795747c7239e30c7c1fd37c82ec310d9273eca36e1013feae0910bb7: Status 404 returned error can't find the container with id bafb6eed795747c7239e30c7c1fd37c82ec310d9273eca36e1013feae0910bb7 Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.414034 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd84d754-4092-4b2c-a8aa-6727c88c5100-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.506133 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.530742 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.541201 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.541735 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-central-agent" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.541761 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-central-agent" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.541785 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="proxy-httpd" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.541794 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="proxy-httpd" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.541806 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-notification-agent" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.541815 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-notification-agent" Dec 04 12:05:43 crc kubenswrapper[4979]: E1204 12:05:43.541833 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="sg-core" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.541840 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="sg-core" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.542062 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="proxy-httpd" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.542098 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-notification-agent" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.542113 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="ceilometer-central-agent" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.542130 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" containerName="sg-core" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.544073 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.548892 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.549262 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.549344 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632142 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632520 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l94v2\" (UniqueName: \"kubernetes.io/projected/5ead262f-a4ca-45bc-a272-8a0648b33402-kube-api-access-l94v2\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632551 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-run-httpd\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632583 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632661 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-log-httpd\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632688 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-scripts\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.632717 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-config-data\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734244 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l94v2\" (UniqueName: \"kubernetes.io/projected/5ead262f-a4ca-45bc-a272-8a0648b33402-kube-api-access-l94v2\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-run-httpd\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-log-httpd\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734444 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-scripts\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734467 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-config-data\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734507 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.734943 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-run-httpd\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.735068 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-log-httpd\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.740162 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.740574 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.741681 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-scripts\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.743860 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-config-data\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.754479 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l94v2\" (UniqueName: \"kubernetes.io/projected/5ead262f-a4ca-45bc-a272-8a0648b33402-kube-api-access-l94v2\") pod \"ceilometer-0\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " pod="openstack/ceilometer-0" Dec 04 12:05:43 crc kubenswrapper[4979]: I1204 12:05:43.863967 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.180216 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4ad703ff-c521-40f1-9fa4-09fca7726079","Type":"ContainerStarted","Data":"07d2a54c7f114041633c7190d149a251778f6bde1cdbf1ccdf26922e6d2aeed8"} Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.180585 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4ad703ff-c521-40f1-9fa4-09fca7726079","Type":"ContainerStarted","Data":"bafb6eed795747c7239e30c7c1fd37c82ec310d9273eca36e1013feae0910bb7"} Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.183270 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c","Type":"ContainerStarted","Data":"063e5850e136541b04a3fbcd523818e0db023f1015cc97b6258d7d2440f0fc84"} Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.200800 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.200780871 podStartE2EDuration="3.200780871s" podCreationTimestamp="2025-12-04 12:05:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:44.198640422 +0000 UTC m=+1368.472936226" watchObservedRunningTime="2025-12-04 12:05:44.200780871 +0000 UTC m=+1368.475076675" Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.216067 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56f16141-154b-401c-bf4e-ca6d7b73c040" path="/var/lib/kubelet/pods/56f16141-154b-401c-bf4e-ca6d7b73c040/volumes" Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.216878 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd84d754-4092-4b2c-a8aa-6727c88c5100" path="/var/lib/kubelet/pods/bd84d754-4092-4b2c-a8aa-6727c88c5100/volumes" Dec 04 12:05:44 crc kubenswrapper[4979]: I1204 12:05:44.326109 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:44 crc kubenswrapper[4979]: W1204 12:05:44.332945 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ead262f_a4ca_45bc_a272_8a0648b33402.slice/crio-3a73163b2af660e3aab4c598df9f1474a45c2e7a5c247c2efb4b5d488c18e552 WatchSource:0}: Error finding container 3a73163b2af660e3aab4c598df9f1474a45c2e7a5c247c2efb4b5d488c18e552: Status 404 returned error can't find the container with id 3a73163b2af660e3aab4c598df9f1474a45c2e7a5c247c2efb4b5d488c18e552 Dec 04 12:05:45 crc kubenswrapper[4979]: I1204 12:05:45.206512 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerStarted","Data":"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3"} Dec 04 12:05:45 crc kubenswrapper[4979]: I1204 12:05:45.207378 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerStarted","Data":"3a73163b2af660e3aab4c598df9f1474a45c2e7a5c247c2efb4b5d488c18e552"} Dec 04 12:05:45 crc kubenswrapper[4979]: I1204 12:05:45.218381 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4ad703ff-c521-40f1-9fa4-09fca7726079","Type":"ContainerStarted","Data":"3d54c5d9cb36dae2f6bcd838a3e0fddaf4c2d1f19e347ee987290f673c8efd37"} Dec 04 12:05:45 crc kubenswrapper[4979]: I1204 12:05:45.218572 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 12:05:45 crc kubenswrapper[4979]: I1204 12:05:45.246385 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.246368469 podStartE2EDuration="3.246368469s" podCreationTimestamp="2025-12-04 12:05:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:45.243514821 +0000 UTC m=+1369.517810635" watchObservedRunningTime="2025-12-04 12:05:45.246368469 +0000 UTC m=+1369.520664273" Dec 04 12:05:46 crc kubenswrapper[4979]: I1204 12:05:46.241814 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerStarted","Data":"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8"} Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.259787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerStarted","Data":"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955"} Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.439009 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.439058 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.475033 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.484670 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.851328 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-ps247"] Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.853824 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.882928 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ps247"] Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.950416 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-zb6xt"] Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.951585 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:47 crc kubenswrapper[4979]: I1204 12:05:47.963972 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zb6xt"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.013976 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8c7vf\" (UniqueName: \"kubernetes.io/projected/295397f3-f237-491c-82d0-780fe406c9e1-kube-api-access-8c7vf\") pod \"nova-api-db-create-ps247\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.014159 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295397f3-f237-491c-82d0-780fe406c9e1-operator-scripts\") pod \"nova-api-db-create-ps247\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.055138 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-9kwrg"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.056360 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.062204 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c904-account-create-update-c2vz5"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.063464 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.066338 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.069619 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9kwrg"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.079956 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c904-account-create-update-c2vz5"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.116782 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8c7vf\" (UniqueName: \"kubernetes.io/projected/295397f3-f237-491c-82d0-780fe406c9e1-kube-api-access-8c7vf\") pod \"nova-api-db-create-ps247\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.116851 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xftbh\" (UniqueName: \"kubernetes.io/projected/b03c2a80-d7a1-4eaf-b571-16a264fdc583-kube-api-access-xftbh\") pod \"nova-cell0-db-create-zb6xt\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.116890 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295397f3-f237-491c-82d0-780fe406c9e1-operator-scripts\") pod \"nova-api-db-create-ps247\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.116907 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03c2a80-d7a1-4eaf-b571-16a264fdc583-operator-scripts\") pod \"nova-cell0-db-create-zb6xt\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.118776 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295397f3-f237-491c-82d0-780fe406c9e1-operator-scripts\") pod \"nova-api-db-create-ps247\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.143570 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8c7vf\" (UniqueName: \"kubernetes.io/projected/295397f3-f237-491c-82d0-780fe406c9e1-kube-api-access-8c7vf\") pod \"nova-api-db-create-ps247\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.199465 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.218491 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltd4w\" (UniqueName: \"kubernetes.io/projected/c5529cf7-2998-44bc-bb8f-3cc8da54450d-kube-api-access-ltd4w\") pod \"nova-cell1-db-create-9kwrg\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.218561 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xftbh\" (UniqueName: \"kubernetes.io/projected/b03c2a80-d7a1-4eaf-b571-16a264fdc583-kube-api-access-xftbh\") pod \"nova-cell0-db-create-zb6xt\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.218603 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tbxs\" (UniqueName: \"kubernetes.io/projected/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-kube-api-access-4tbxs\") pod \"nova-api-c904-account-create-update-c2vz5\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.218633 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03c2a80-d7a1-4eaf-b571-16a264fdc583-operator-scripts\") pod \"nova-cell0-db-create-zb6xt\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.218716 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-operator-scripts\") pod \"nova-api-c904-account-create-update-c2vz5\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.218745 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5529cf7-2998-44bc-bb8f-3cc8da54450d-operator-scripts\") pod \"nova-cell1-db-create-9kwrg\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.219674 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03c2a80-d7a1-4eaf-b571-16a264fdc583-operator-scripts\") pod \"nova-cell0-db-create-zb6xt\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.240941 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xftbh\" (UniqueName: \"kubernetes.io/projected/b03c2a80-d7a1-4eaf-b571-16a264fdc583-kube-api-access-xftbh\") pod \"nova-cell0-db-create-zb6xt\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.276021 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.283454 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerStarted","Data":"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7"} Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.284183 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.284223 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.301836 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-2b03-account-create-update-d44w9"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.303173 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.306474 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2b03-account-create-update-d44w9"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.309839 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.323445 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltd4w\" (UniqueName: \"kubernetes.io/projected/c5529cf7-2998-44bc-bb8f-3cc8da54450d-kube-api-access-ltd4w\") pod \"nova-cell1-db-create-9kwrg\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.323554 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tbxs\" (UniqueName: \"kubernetes.io/projected/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-kube-api-access-4tbxs\") pod \"nova-api-c904-account-create-update-c2vz5\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.323694 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-operator-scripts\") pod \"nova-api-c904-account-create-update-c2vz5\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.323736 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5529cf7-2998-44bc-bb8f-3cc8da54450d-operator-scripts\") pod \"nova-cell1-db-create-9kwrg\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.324887 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5529cf7-2998-44bc-bb8f-3cc8da54450d-operator-scripts\") pod \"nova-cell1-db-create-9kwrg\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.327161 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-operator-scripts\") pod \"nova-api-c904-account-create-update-c2vz5\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.354946 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltd4w\" (UniqueName: \"kubernetes.io/projected/c5529cf7-2998-44bc-bb8f-3cc8da54450d-kube-api-access-ltd4w\") pod \"nova-cell1-db-create-9kwrg\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.373983 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tbxs\" (UniqueName: \"kubernetes.io/projected/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-kube-api-access-4tbxs\") pod \"nova-api-c904-account-create-update-c2vz5\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.376757 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.384129 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.407178932 podStartE2EDuration="5.384102872s" podCreationTimestamp="2025-12-04 12:05:43 +0000 UTC" firstStartedPulling="2025-12-04 12:05:44.335861283 +0000 UTC m=+1368.610157087" lastFinishedPulling="2025-12-04 12:05:47.312785213 +0000 UTC m=+1371.587081027" observedRunningTime="2025-12-04 12:05:48.340761733 +0000 UTC m=+1372.615057537" watchObservedRunningTime="2025-12-04 12:05:48.384102872 +0000 UTC m=+1372.658398676" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.384888 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.418349 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-71eb-account-create-update-wdcd2"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.419434 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.421829 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.426526 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsjrv\" (UniqueName: \"kubernetes.io/projected/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-kube-api-access-rsjrv\") pod \"nova-cell0-2b03-account-create-update-d44w9\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.426653 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-operator-scripts\") pod \"nova-cell0-2b03-account-create-update-d44w9\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.454442 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-71eb-account-create-update-wdcd2"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.529166 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgbq9\" (UniqueName: \"kubernetes.io/projected/113cba3a-dbbb-44a6-8342-3e287e77cb4d-kube-api-access-lgbq9\") pod \"nova-cell1-71eb-account-create-update-wdcd2\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.529268 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-operator-scripts\") pod \"nova-cell0-2b03-account-create-update-d44w9\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.529463 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/113cba3a-dbbb-44a6-8342-3e287e77cb4d-operator-scripts\") pod \"nova-cell1-71eb-account-create-update-wdcd2\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.529482 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsjrv\" (UniqueName: \"kubernetes.io/projected/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-kube-api-access-rsjrv\") pod \"nova-cell0-2b03-account-create-update-d44w9\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.529974 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-operator-scripts\") pod \"nova-cell0-2b03-account-create-update-d44w9\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.548966 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsjrv\" (UniqueName: \"kubernetes.io/projected/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-kube-api-access-rsjrv\") pod \"nova-cell0-2b03-account-create-update-d44w9\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.631469 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/113cba3a-dbbb-44a6-8342-3e287e77cb4d-operator-scripts\") pod \"nova-cell1-71eb-account-create-update-wdcd2\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.631592 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgbq9\" (UniqueName: \"kubernetes.io/projected/113cba3a-dbbb-44a6-8342-3e287e77cb4d-kube-api-access-lgbq9\") pod \"nova-cell1-71eb-account-create-update-wdcd2\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.632598 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/113cba3a-dbbb-44a6-8342-3e287e77cb4d-operator-scripts\") pod \"nova-cell1-71eb-account-create-update-wdcd2\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.655827 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgbq9\" (UniqueName: \"kubernetes.io/projected/113cba3a-dbbb-44a6-8342-3e287e77cb4d-kube-api-access-lgbq9\") pod \"nova-cell1-71eb-account-create-update-wdcd2\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.744140 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-ps247"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.834117 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.896425 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.991799 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c904-account-create-update-c2vz5"] Dec 04 12:05:48 crc kubenswrapper[4979]: I1204 12:05:48.999927 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zb6xt"] Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.111594 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-9kwrg"] Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.229179 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-71eb-account-create-update-wdcd2"] Dec 04 12:05:49 crc kubenswrapper[4979]: W1204 12:05:49.244630 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod113cba3a_dbbb_44a6_8342_3e287e77cb4d.slice/crio-7d46efd727c0e8d613d36ac99e3fc6e0f5e7dceb3baeecf1372561af36a76c6b WatchSource:0}: Error finding container 7d46efd727c0e8d613d36ac99e3fc6e0f5e7dceb3baeecf1372561af36a76c6b: Status 404 returned error can't find the container with id 7d46efd727c0e8d613d36ac99e3fc6e0f5e7dceb3baeecf1372561af36a76c6b Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.298274 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zb6xt" event={"ID":"b03c2a80-d7a1-4eaf-b571-16a264fdc583","Type":"ContainerStarted","Data":"781c44b343a8b542a2e92441044c2ef19e45c36eb72c633006adb84d69aeb4c7"} Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.304379 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kwrg" event={"ID":"c5529cf7-2998-44bc-bb8f-3cc8da54450d","Type":"ContainerStarted","Data":"751ca9c73b63f5df0065ba9058c3b80a1e5ac6ec4db7218bf719259ab092597d"} Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.306978 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c904-account-create-update-c2vz5" event={"ID":"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1","Type":"ContainerStarted","Data":"f20ab65be06258af0810b2b7933ea2c727f9b87375cd7d83814f0e5d6d9411b8"} Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.317653 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ps247" event={"ID":"295397f3-f237-491c-82d0-780fe406c9e1","Type":"ContainerStarted","Data":"d03c505af873328ff9c508812d00120785b8d279d1fa8b7860baaf6741c42a6c"} Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.317715 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ps247" event={"ID":"295397f3-f237-491c-82d0-780fe406c9e1","Type":"ContainerStarted","Data":"9cc041a113aa892c97dc52fedcd76604b8097099790efcf299fab53b62afdba6"} Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.321844 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" event={"ID":"113cba3a-dbbb-44a6-8342-3e287e77cb4d","Type":"ContainerStarted","Data":"7d46efd727c0e8d613d36ac99e3fc6e0f5e7dceb3baeecf1372561af36a76c6b"} Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.322824 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.324955 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-2b03-account-create-update-d44w9"] Dec 04 12:05:49 crc kubenswrapper[4979]: W1204 12:05:49.335861 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfac7d10c_b3d4_4fbe_8b09_939316b41b8d.slice/crio-c9f827dec149464b37b835f089456c47f42c4b466a158d52add4fd33ba4f4ac3 WatchSource:0}: Error finding container c9f827dec149464b37b835f089456c47f42c4b466a158d52add4fd33ba4f4ac3: Status 404 returned error can't find the container with id c9f827dec149464b37b835f089456c47f42c4b466a158d52add4fd33ba4f4ac3 Dec 04 12:05:49 crc kubenswrapper[4979]: I1204 12:05:49.344816 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-ps247" podStartSLOduration=2.344793313 podStartE2EDuration="2.344793313s" podCreationTimestamp="2025-12-04 12:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:05:49.335439208 +0000 UTC m=+1373.609735012" watchObservedRunningTime="2025-12-04 12:05:49.344793313 +0000 UTC m=+1373.619089117" Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.175779 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.332252 4979 generic.go:334] "Generic (PLEG): container finished" podID="562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" containerID="1e209cfbc1ecb28128f57fb249b4b4a7d088a53a2531b190fcf5e7e253d5aed3" exitCode=0 Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.332321 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c904-account-create-update-c2vz5" event={"ID":"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1","Type":"ContainerDied","Data":"1e209cfbc1ecb28128f57fb249b4b4a7d088a53a2531b190fcf5e7e253d5aed3"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.333978 4979 generic.go:334] "Generic (PLEG): container finished" podID="295397f3-f237-491c-82d0-780fe406c9e1" containerID="d03c505af873328ff9c508812d00120785b8d279d1fa8b7860baaf6741c42a6c" exitCode=0 Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.334043 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ps247" event={"ID":"295397f3-f237-491c-82d0-780fe406c9e1","Type":"ContainerDied","Data":"d03c505af873328ff9c508812d00120785b8d279d1fa8b7860baaf6741c42a6c"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.335662 4979 generic.go:334] "Generic (PLEG): container finished" podID="fac7d10c-b3d4-4fbe-8b09-939316b41b8d" containerID="9acd8b6826bd47b62b1e082bc092cd3865472726e361e5dceb1e5f9d607411ba" exitCode=0 Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.335716 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" event={"ID":"fac7d10c-b3d4-4fbe-8b09-939316b41b8d","Type":"ContainerDied","Data":"9acd8b6826bd47b62b1e082bc092cd3865472726e361e5dceb1e5f9d607411ba"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.335735 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" event={"ID":"fac7d10c-b3d4-4fbe-8b09-939316b41b8d","Type":"ContainerStarted","Data":"c9f827dec149464b37b835f089456c47f42c4b466a158d52add4fd33ba4f4ac3"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.337307 4979 generic.go:334] "Generic (PLEG): container finished" podID="113cba3a-dbbb-44a6-8342-3e287e77cb4d" containerID="453fe0a328e8492eb90c3463bc6db1ca1052fc0f0ce8b9544f2ae3008198ff55" exitCode=0 Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.337387 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" event={"ID":"113cba3a-dbbb-44a6-8342-3e287e77cb4d","Type":"ContainerDied","Data":"453fe0a328e8492eb90c3463bc6db1ca1052fc0f0ce8b9544f2ae3008198ff55"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.338966 4979 generic.go:334] "Generic (PLEG): container finished" podID="b03c2a80-d7a1-4eaf-b571-16a264fdc583" containerID="a07ddf9ea78efef5f6585f245d44e118f0a95206043e0fd406aef1c64f4cd3ca" exitCode=0 Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.339022 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zb6xt" event={"ID":"b03c2a80-d7a1-4eaf-b571-16a264fdc583","Type":"ContainerDied","Data":"a07ddf9ea78efef5f6585f245d44e118f0a95206043e0fd406aef1c64f4cd3ca"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.340603 4979 generic.go:334] "Generic (PLEG): container finished" podID="c5529cf7-2998-44bc-bb8f-3cc8da54450d" containerID="bf93f510653a27db045fa5f965a3176f65591f23d896c18d1be176054af8b89f" exitCode=0 Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.340660 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kwrg" event={"ID":"c5529cf7-2998-44bc-bb8f-3cc8da54450d","Type":"ContainerDied","Data":"bf93f510653a27db045fa5f965a3176f65591f23d896c18d1be176054af8b89f"} Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.340712 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.340726 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.482972 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:50 crc kubenswrapper[4979]: I1204 12:05:50.641000 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.355728 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-central-agent" containerID="cri-o://6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" gracePeriod=30 Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.357343 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="proxy-httpd" containerID="cri-o://bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" gracePeriod=30 Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.357744 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-notification-agent" containerID="cri-o://669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" gracePeriod=30 Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.358517 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="sg-core" containerID="cri-o://1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" gracePeriod=30 Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.478776 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.480017 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.549979 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.562212 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.805288 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.903339 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgbq9\" (UniqueName: \"kubernetes.io/projected/113cba3a-dbbb-44a6-8342-3e287e77cb4d-kube-api-access-lgbq9\") pod \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.903476 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/113cba3a-dbbb-44a6-8342-3e287e77cb4d-operator-scripts\") pod \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\" (UID: \"113cba3a-dbbb-44a6-8342-3e287e77cb4d\") " Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.905754 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/113cba3a-dbbb-44a6-8342-3e287e77cb4d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "113cba3a-dbbb-44a6-8342-3e287e77cb4d" (UID: "113cba3a-dbbb-44a6-8342-3e287e77cb4d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.924575 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/113cba3a-dbbb-44a6-8342-3e287e77cb4d-kube-api-access-lgbq9" (OuterVolumeSpecName: "kube-api-access-lgbq9") pod "113cba3a-dbbb-44a6-8342-3e287e77cb4d" (UID: "113cba3a-dbbb-44a6-8342-3e287e77cb4d"). InnerVolumeSpecName "kube-api-access-lgbq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.962838 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.991847 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:51 crc kubenswrapper[4979]: I1204 12:05:51.992707 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.002115 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.007520 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgbq9\" (UniqueName: \"kubernetes.io/projected/113cba3a-dbbb-44a6-8342-3e287e77cb4d-kube-api-access-lgbq9\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.007550 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/113cba3a-dbbb-44a6-8342-3e287e77cb4d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.008675 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108200 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5529cf7-2998-44bc-bb8f-3cc8da54450d-operator-scripts\") pod \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108259 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03c2a80-d7a1-4eaf-b571-16a264fdc583-operator-scripts\") pod \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108426 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8c7vf\" (UniqueName: \"kubernetes.io/projected/295397f3-f237-491c-82d0-780fe406c9e1-kube-api-access-8c7vf\") pod \"295397f3-f237-491c-82d0-780fe406c9e1\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108473 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltd4w\" (UniqueName: \"kubernetes.io/projected/c5529cf7-2998-44bc-bb8f-3cc8da54450d-kube-api-access-ltd4w\") pod \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\" (UID: \"c5529cf7-2998-44bc-bb8f-3cc8da54450d\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108515 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-operator-scripts\") pod \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108558 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsjrv\" (UniqueName: \"kubernetes.io/projected/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-kube-api-access-rsjrv\") pod \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\" (UID: \"fac7d10c-b3d4-4fbe-8b09-939316b41b8d\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108613 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295397f3-f237-491c-82d0-780fe406c9e1-operator-scripts\") pod \"295397f3-f237-491c-82d0-780fe406c9e1\" (UID: \"295397f3-f237-491c-82d0-780fe406c9e1\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108653 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4tbxs\" (UniqueName: \"kubernetes.io/projected/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-kube-api-access-4tbxs\") pod \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108655 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b03c2a80-d7a1-4eaf-b571-16a264fdc583-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b03c2a80-d7a1-4eaf-b571-16a264fdc583" (UID: "b03c2a80-d7a1-4eaf-b571-16a264fdc583"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108703 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5529cf7-2998-44bc-bb8f-3cc8da54450d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5529cf7-2998-44bc-bb8f-3cc8da54450d" (UID: "c5529cf7-2998-44bc-bb8f-3cc8da54450d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108727 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-operator-scripts\") pod \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\" (UID: \"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108754 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xftbh\" (UniqueName: \"kubernetes.io/projected/b03c2a80-d7a1-4eaf-b571-16a264fdc583-kube-api-access-xftbh\") pod \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\" (UID: \"b03c2a80-d7a1-4eaf-b571-16a264fdc583\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.108927 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fac7d10c-b3d4-4fbe-8b09-939316b41b8d" (UID: "fac7d10c-b3d4-4fbe-8b09-939316b41b8d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.109074 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/295397f3-f237-491c-82d0-780fe406c9e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "295397f3-f237-491c-82d0-780fe406c9e1" (UID: "295397f3-f237-491c-82d0-780fe406c9e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.109230 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5529cf7-2998-44bc-bb8f-3cc8da54450d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.109551 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b03c2a80-d7a1-4eaf-b571-16a264fdc583-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.109565 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.109577 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/295397f3-f237-491c-82d0-780fe406c9e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.109797 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" (UID: "562eca34-0bf6-4bc9-a2f7-9bf2348b78c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.112436 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-kube-api-access-4tbxs" (OuterVolumeSpecName: "kube-api-access-4tbxs") pod "562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" (UID: "562eca34-0bf6-4bc9-a2f7-9bf2348b78c1"). InnerVolumeSpecName "kube-api-access-4tbxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.112893 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03c2a80-d7a1-4eaf-b571-16a264fdc583-kube-api-access-xftbh" (OuterVolumeSpecName: "kube-api-access-xftbh") pod "b03c2a80-d7a1-4eaf-b571-16a264fdc583" (UID: "b03c2a80-d7a1-4eaf-b571-16a264fdc583"). InnerVolumeSpecName "kube-api-access-xftbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.114694 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5529cf7-2998-44bc-bb8f-3cc8da54450d-kube-api-access-ltd4w" (OuterVolumeSpecName: "kube-api-access-ltd4w") pod "c5529cf7-2998-44bc-bb8f-3cc8da54450d" (UID: "c5529cf7-2998-44bc-bb8f-3cc8da54450d"). InnerVolumeSpecName "kube-api-access-ltd4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.114752 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-kube-api-access-rsjrv" (OuterVolumeSpecName: "kube-api-access-rsjrv") pod "fac7d10c-b3d4-4fbe-8b09-939316b41b8d" (UID: "fac7d10c-b3d4-4fbe-8b09-939316b41b8d"). InnerVolumeSpecName "kube-api-access-rsjrv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.115727 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/295397f3-f237-491c-82d0-780fe406c9e1-kube-api-access-8c7vf" (OuterVolumeSpecName: "kube-api-access-8c7vf") pod "295397f3-f237-491c-82d0-780fe406c9e1" (UID: "295397f3-f237-491c-82d0-780fe406c9e1"). InnerVolumeSpecName "kube-api-access-8c7vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.210818 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4tbxs\" (UniqueName: \"kubernetes.io/projected/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-kube-api-access-4tbxs\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.211271 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.211289 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xftbh\" (UniqueName: \"kubernetes.io/projected/b03c2a80-d7a1-4eaf-b571-16a264fdc583-kube-api-access-xftbh\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.211309 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8c7vf\" (UniqueName: \"kubernetes.io/projected/295397f3-f237-491c-82d0-780fe406c9e1-kube-api-access-8c7vf\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.211318 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltd4w\" (UniqueName: \"kubernetes.io/projected/c5529cf7-2998-44bc-bb8f-3cc8da54450d-kube-api-access-ltd4w\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.211328 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsjrv\" (UniqueName: \"kubernetes.io/projected/fac7d10c-b3d4-4fbe-8b09-939316b41b8d-kube-api-access-rsjrv\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.276620 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.369761 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" event={"ID":"fac7d10c-b3d4-4fbe-8b09-939316b41b8d","Type":"ContainerDied","Data":"c9f827dec149464b37b835f089456c47f42c4b466a158d52add4fd33ba4f4ac3"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.369826 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9f827dec149464b37b835f089456c47f42c4b466a158d52add4fd33ba4f4ac3" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.369780 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-2b03-account-create-update-d44w9" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.371353 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" event={"ID":"113cba3a-dbbb-44a6-8342-3e287e77cb4d","Type":"ContainerDied","Data":"7d46efd727c0e8d613d36ac99e3fc6e0f5e7dceb3baeecf1372561af36a76c6b"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.371400 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d46efd727c0e8d613d36ac99e3fc6e0f5e7dceb3baeecf1372561af36a76c6b" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.371446 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-71eb-account-create-update-wdcd2" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.373498 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zb6xt" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.374049 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zb6xt" event={"ID":"b03c2a80-d7a1-4eaf-b571-16a264fdc583","Type":"ContainerDied","Data":"781c44b343a8b542a2e92441044c2ef19e45c36eb72c633006adb84d69aeb4c7"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.374088 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="781c44b343a8b542a2e92441044c2ef19e45c36eb72c633006adb84d69aeb4c7" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.380240 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-9kwrg" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.380368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-9kwrg" event={"ID":"c5529cf7-2998-44bc-bb8f-3cc8da54450d","Type":"ContainerDied","Data":"751ca9c73b63f5df0065ba9058c3b80a1e5ac6ec4db7218bf719259ab092597d"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.380409 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="751ca9c73b63f5df0065ba9058c3b80a1e5ac6ec4db7218bf719259ab092597d" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.382643 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c904-account-create-update-c2vz5" event={"ID":"562eca34-0bf6-4bc9-a2f7-9bf2348b78c1","Type":"ContainerDied","Data":"f20ab65be06258af0810b2b7933ea2c727f9b87375cd7d83814f0e5d6d9411b8"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.382677 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f20ab65be06258af0810b2b7933ea2c727f9b87375cd7d83814f0e5d6d9411b8" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.382723 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c904-account-create-update-c2vz5" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386012 4979 generic.go:334] "Generic (PLEG): container finished" podID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" exitCode=0 Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386046 4979 generic.go:334] "Generic (PLEG): container finished" podID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" exitCode=2 Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386059 4979 generic.go:334] "Generic (PLEG): container finished" podID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" exitCode=0 Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386071 4979 generic.go:334] "Generic (PLEG): container finished" podID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" exitCode=0 Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386119 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386130 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerDied","Data":"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386159 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerDied","Data":"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386176 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerDied","Data":"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386189 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerDied","Data":"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386204 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5ead262f-a4ca-45bc-a272-8a0648b33402","Type":"ContainerDied","Data":"3a73163b2af660e3aab4c598df9f1474a45c2e7a5c247c2efb4b5d488c18e552"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.386222 4979 scope.go:117] "RemoveContainer" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.388387 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-ps247" event={"ID":"295397f3-f237-491c-82d0-780fe406c9e1","Type":"ContainerDied","Data":"9cc041a113aa892c97dc52fedcd76604b8097099790efcf299fab53b62afdba6"} Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.388426 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cc041a113aa892c97dc52fedcd76604b8097099790efcf299fab53b62afdba6" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.388604 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-ps247" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.388647 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.389251 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.406641 4979 scope.go:117] "RemoveContainer" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.423935 4979 scope.go:117] "RemoveContainer" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.427433 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-log-httpd\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.427524 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-scripts\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.427699 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-sg-core-conf-yaml\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.427776 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-config-data\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.427943 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-combined-ca-bundle\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.428015 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l94v2\" (UniqueName: \"kubernetes.io/projected/5ead262f-a4ca-45bc-a272-8a0648b33402-kube-api-access-l94v2\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.428076 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-run-httpd\") pod \"5ead262f-a4ca-45bc-a272-8a0648b33402\" (UID: \"5ead262f-a4ca-45bc-a272-8a0648b33402\") " Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.428593 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.428779 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.429197 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.431993 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-scripts" (OuterVolumeSpecName: "scripts") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.433054 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ead262f-a4ca-45bc-a272-8a0648b33402-kube-api-access-l94v2" (OuterVolumeSpecName: "kube-api-access-l94v2") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "kube-api-access-l94v2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.453056 4979 scope.go:117] "RemoveContainer" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.461503 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.475222 4979 scope.go:117] "RemoveContainer" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.476121 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": container with ID starting with bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7 not found: ID does not exist" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.476156 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7"} err="failed to get container status \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": rpc error: code = NotFound desc = could not find container \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": container with ID starting with bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.476176 4979 scope.go:117] "RemoveContainer" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.476552 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": container with ID starting with 1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955 not found: ID does not exist" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.476607 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955"} err="failed to get container status \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": rpc error: code = NotFound desc = could not find container \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": container with ID starting with 1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.476622 4979 scope.go:117] "RemoveContainer" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.476977 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": container with ID starting with 669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8 not found: ID does not exist" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.477026 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8"} err="failed to get container status \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": rpc error: code = NotFound desc = could not find container \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": container with ID starting with 669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.477046 4979 scope.go:117] "RemoveContainer" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.477363 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": container with ID starting with 6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3 not found: ID does not exist" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.477402 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3"} err="failed to get container status \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": rpc error: code = NotFound desc = could not find container \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": container with ID starting with 6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.477428 4979 scope.go:117] "RemoveContainer" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.478346 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7"} err="failed to get container status \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": rpc error: code = NotFound desc = could not find container \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": container with ID starting with bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.478418 4979 scope.go:117] "RemoveContainer" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.478723 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955"} err="failed to get container status \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": rpc error: code = NotFound desc = could not find container \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": container with ID starting with 1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.478743 4979 scope.go:117] "RemoveContainer" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.479047 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8"} err="failed to get container status \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": rpc error: code = NotFound desc = could not find container \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": container with ID starting with 669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.479087 4979 scope.go:117] "RemoveContainer" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.479433 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3"} err="failed to get container status \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": rpc error: code = NotFound desc = could not find container \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": container with ID starting with 6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.479463 4979 scope.go:117] "RemoveContainer" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.480670 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7"} err="failed to get container status \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": rpc error: code = NotFound desc = could not find container \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": container with ID starting with bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.480690 4979 scope.go:117] "RemoveContainer" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.480936 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955"} err="failed to get container status \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": rpc error: code = NotFound desc = could not find container \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": container with ID starting with 1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.481000 4979 scope.go:117] "RemoveContainer" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.481266 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8"} err="failed to get container status \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": rpc error: code = NotFound desc = could not find container \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": container with ID starting with 669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.481288 4979 scope.go:117] "RemoveContainer" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.481582 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3"} err="failed to get container status \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": rpc error: code = NotFound desc = could not find container \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": container with ID starting with 6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.481601 4979 scope.go:117] "RemoveContainer" containerID="bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.484956 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7"} err="failed to get container status \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": rpc error: code = NotFound desc = could not find container \"bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7\": container with ID starting with bc8d6a67a37637cfe3487fb94823bc232f4bde26bb480e96a24f08634ad06bd7 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.485009 4979 scope.go:117] "RemoveContainer" containerID="1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.485440 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955"} err="failed to get container status \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": rpc error: code = NotFound desc = could not find container \"1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955\": container with ID starting with 1c33145e718910366a1e607e85ee9e720395eec11c0093e888a5ed3c42447955 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.485474 4979 scope.go:117] "RemoveContainer" containerID="669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.486758 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8"} err="failed to get container status \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": rpc error: code = NotFound desc = could not find container \"669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8\": container with ID starting with 669b49fed5c1dfab2ed773a405ffccab49cf7c68ba313d0cba7cc13262f390f8 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.486782 4979 scope.go:117] "RemoveContainer" containerID="6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.487014 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3"} err="failed to get container status \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": rpc error: code = NotFound desc = could not find container \"6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3\": container with ID starting with 6a8ee6d9e42000ce34591f8a20e2a38034f5474a52ae3cd7da10ea183f23e4c3 not found: ID does not exist" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.505384 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.531130 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.531159 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.531168 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.531177 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l94v2\" (UniqueName: \"kubernetes.io/projected/5ead262f-a4ca-45bc-a272-8a0648b33402-kube-api-access-l94v2\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.531186 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5ead262f-a4ca-45bc-a272-8a0648b33402-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.539242 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-config-data" (OuterVolumeSpecName: "config-data") pod "5ead262f-a4ca-45bc-a272-8a0648b33402" (UID: "5ead262f-a4ca-45bc-a272-8a0648b33402"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.633149 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ead262f-a4ca-45bc-a272-8a0648b33402-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.732234 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.742337 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.748865 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749224 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5529cf7-2998-44bc-bb8f-3cc8da54450d" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749243 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5529cf7-2998-44bc-bb8f-3cc8da54450d" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749258 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="proxy-httpd" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749265 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="proxy-httpd" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749279 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-central-agent" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749286 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-central-agent" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749312 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749319 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749327 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-notification-agent" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749333 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-notification-agent" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749347 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="295397f3-f237-491c-82d0-780fe406c9e1" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749353 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="295397f3-f237-491c-82d0-780fe406c9e1" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749364 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="sg-core" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749370 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="sg-core" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749383 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="113cba3a-dbbb-44a6-8342-3e287e77cb4d" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749389 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="113cba3a-dbbb-44a6-8342-3e287e77cb4d" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749399 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac7d10c-b3d4-4fbe-8b09-939316b41b8d" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749406 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac7d10c-b3d4-4fbe-8b09-939316b41b8d" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: E1204 12:05:52.749417 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03c2a80-d7a1-4eaf-b571-16a264fdc583" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749423 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03c2a80-d7a1-4eaf-b571-16a264fdc583" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749576 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03c2a80-d7a1-4eaf-b571-16a264fdc583" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749589 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="proxy-httpd" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749601 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac7d10c-b3d4-4fbe-8b09-939316b41b8d" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749618 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="295397f3-f237-491c-82d0-780fe406c9e1" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749627 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-central-agent" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749638 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="113cba3a-dbbb-44a6-8342-3e287e77cb4d" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749645 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="sg-core" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749655 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" containerName="ceilometer-notification-agent" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749661 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" containerName="mariadb-account-create-update" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.749706 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5529cf7-2998-44bc-bb8f-3cc8da54450d" containerName="mariadb-database-create" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.752225 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.757432 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.757635 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.789850 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.843759 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qw2k\" (UniqueName: \"kubernetes.io/projected/a33c654a-91bc-4019-ac22-b255cb307f9b-kube-api-access-5qw2k\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.843823 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-config-data\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.843859 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-scripts\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.843920 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-run-httpd\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.844029 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-log-httpd\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.844069 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.844286 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946261 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qw2k\" (UniqueName: \"kubernetes.io/projected/a33c654a-91bc-4019-ac22-b255cb307f9b-kube-api-access-5qw2k\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946386 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-config-data\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-scripts\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946458 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-run-httpd\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946520 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-log-httpd\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946551 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.946710 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.947162 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-log-httpd\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.947325 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-run-httpd\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.952531 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.962607 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.963413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-scripts\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.963795 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-config-data\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:52 crc kubenswrapper[4979]: I1204 12:05:52.971107 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qw2k\" (UniqueName: \"kubernetes.io/projected/a33c654a-91bc-4019-ac22-b255cb307f9b-kube-api-access-5qw2k\") pod \"ceilometer-0\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " pod="openstack/ceilometer-0" Dec 04 12:05:53 crc kubenswrapper[4979]: I1204 12:05:53.105737 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:05:53 crc kubenswrapper[4979]: I1204 12:05:53.660069 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.208055 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ead262f-a4ca-45bc-a272-8a0648b33402" path="/var/lib/kubelet/pods/5ead262f-a4ca-45bc-a272-8a0648b33402/volumes" Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.413650 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.413675 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.414425 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerStarted","Data":"d1787d30e93aaabf4110bd689fcff1607fabe56138f512c93f3a9759e42bb14d"} Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.414457 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerStarted","Data":"351e7103733539e18e99b929bb713c1525d4c3b0297bbac75aaa67b4516fad98"} Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.662428 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.741257 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.772237 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 12:05:54 crc kubenswrapper[4979]: I1204 12:05:54.868411 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 12:05:55 crc kubenswrapper[4979]: I1204 12:05:55.421277 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerStarted","Data":"9ba3cd701b799904adceb633b48fe34d2978bdc724cd24f658c7b095673130cb"} Dec 04 12:05:56 crc kubenswrapper[4979]: I1204 12:05:56.435481 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerStarted","Data":"0feb0ba424cc089f915437d27ddadc002d0de23fb58a5ff8475542fdcc9bc7a5"} Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.445345 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerStarted","Data":"f90b541672027c1cbbccae80e9c15acdaa39714638e1bad8203cd07f2d5deaf6"} Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.445548 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="sg-core" containerID="cri-o://0feb0ba424cc089f915437d27ddadc002d0de23fb58a5ff8475542fdcc9bc7a5" gracePeriod=30 Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.445606 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.445530 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-central-agent" containerID="cri-o://d1787d30e93aaabf4110bd689fcff1607fabe56138f512c93f3a9759e42bb14d" gracePeriod=30 Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.445583 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-notification-agent" containerID="cri-o://9ba3cd701b799904adceb633b48fe34d2978bdc724cd24f658c7b095673130cb" gracePeriod=30 Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.445550 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="proxy-httpd" containerID="cri-o://f90b541672027c1cbbccae80e9c15acdaa39714638e1bad8203cd07f2d5deaf6" gracePeriod=30 Dec 04 12:05:57 crc kubenswrapper[4979]: I1204 12:05:57.469892 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.437011246 podStartE2EDuration="5.469875438s" podCreationTimestamp="2025-12-04 12:05:52 +0000 UTC" firstStartedPulling="2025-12-04 12:05:53.651269632 +0000 UTC m=+1377.925565436" lastFinishedPulling="2025-12-04 12:05:56.684133824 +0000 UTC m=+1380.958429628" observedRunningTime="2025-12-04 12:05:57.467458412 +0000 UTC m=+1381.741754216" watchObservedRunningTime="2025-12-04 12:05:57.469875438 +0000 UTC m=+1381.744171242" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.249271 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k59t5"] Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.251655 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.254836 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vh9gl" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.254864 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.255777 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.262503 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k59t5"] Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.349947 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.350009 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-config-data\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.350084 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-scripts\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.350132 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv5pl\" (UniqueName: \"kubernetes.io/projected/7b62e2c4-065e-48b2-abc0-03f445f8d787-kube-api-access-fv5pl\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.451164 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-scripts\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.451259 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv5pl\" (UniqueName: \"kubernetes.io/projected/7b62e2c4-065e-48b2-abc0-03f445f8d787-kube-api-access-fv5pl\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.451349 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.451391 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-config-data\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.458742 4979 generic.go:334] "Generic (PLEG): container finished" podID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerID="f90b541672027c1cbbccae80e9c15acdaa39714638e1bad8203cd07f2d5deaf6" exitCode=0 Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.458774 4979 generic.go:334] "Generic (PLEG): container finished" podID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerID="0feb0ba424cc089f915437d27ddadc002d0de23fb58a5ff8475542fdcc9bc7a5" exitCode=2 Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.458784 4979 generic.go:334] "Generic (PLEG): container finished" podID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerID="9ba3cd701b799904adceb633b48fe34d2978bdc724cd24f658c7b095673130cb" exitCode=0 Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.458802 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerDied","Data":"f90b541672027c1cbbccae80e9c15acdaa39714638e1bad8203cd07f2d5deaf6"} Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.458827 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerDied","Data":"0feb0ba424cc089f915437d27ddadc002d0de23fb58a5ff8475542fdcc9bc7a5"} Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.458837 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerDied","Data":"9ba3cd701b799904adceb633b48fe34d2978bdc724cd24f658c7b095673130cb"} Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.459196 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-scripts\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.471030 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv5pl\" (UniqueName: \"kubernetes.io/projected/7b62e2c4-065e-48b2-abc0-03f445f8d787-kube-api-access-fv5pl\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.471999 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-config-data\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.472911 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-k59t5\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:58 crc kubenswrapper[4979]: I1204 12:05:58.575773 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:05:59 crc kubenswrapper[4979]: I1204 12:05:59.073383 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k59t5"] Dec 04 12:05:59 crc kubenswrapper[4979]: W1204 12:05:59.078487 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b62e2c4_065e_48b2_abc0_03f445f8d787.slice/crio-efec5541d7ec816096769b36b43f21ef73ab6747bfc2c9b94638432440942b93 WatchSource:0}: Error finding container efec5541d7ec816096769b36b43f21ef73ab6747bfc2c9b94638432440942b93: Status 404 returned error can't find the container with id efec5541d7ec816096769b36b43f21ef73ab6747bfc2c9b94638432440942b93 Dec 04 12:05:59 crc kubenswrapper[4979]: I1204 12:05:59.470388 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k59t5" event={"ID":"7b62e2c4-065e-48b2-abc0-03f445f8d787","Type":"ContainerStarted","Data":"efec5541d7ec816096769b36b43f21ef73ab6747bfc2c9b94638432440942b93"} Dec 04 12:06:05 crc kubenswrapper[4979]: I1204 12:06:05.978859 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="c067c12b-c393-4f9e-835c-dc2cf68ca81d" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.160:3000/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 12:06:06 crc kubenswrapper[4979]: I1204 12:06:06.548750 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k59t5" event={"ID":"7b62e2c4-065e-48b2-abc0-03f445f8d787","Type":"ContainerStarted","Data":"d6995eed3a9022adb589d900dcb50f240dcc437b7005c2eca54fd004c5fbffa7"} Dec 04 12:06:06 crc kubenswrapper[4979]: I1204 12:06:06.574505 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-k59t5" podStartSLOduration=1.5088103560000001 podStartE2EDuration="8.574485686s" podCreationTimestamp="2025-12-04 12:05:58 +0000 UTC" firstStartedPulling="2025-12-04 12:05:59.080462849 +0000 UTC m=+1383.354758663" lastFinishedPulling="2025-12-04 12:06:06.146138189 +0000 UTC m=+1390.420433993" observedRunningTime="2025-12-04 12:06:06.570807606 +0000 UTC m=+1390.845103430" watchObservedRunningTime="2025-12-04 12:06:06.574485686 +0000 UTC m=+1390.848781490" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.575987 4979 generic.go:334] "Generic (PLEG): container finished" podID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerID="d1787d30e93aaabf4110bd689fcff1607fabe56138f512c93f3a9759e42bb14d" exitCode=0 Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.576586 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerDied","Data":"d1787d30e93aaabf4110bd689fcff1607fabe56138f512c93f3a9759e42bb14d"} Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.687857 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.882945 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-combined-ca-bundle\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.883416 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-run-httpd\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.883543 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qw2k\" (UniqueName: \"kubernetes.io/projected/a33c654a-91bc-4019-ac22-b255cb307f9b-kube-api-access-5qw2k\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.884127 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.884439 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-log-httpd\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.884533 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-scripts\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.884602 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-sg-core-conf-yaml\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.884677 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-config-data\") pod \"a33c654a-91bc-4019-ac22-b255cb307f9b\" (UID: \"a33c654a-91bc-4019-ac22-b255cb307f9b\") " Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.884915 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.885500 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.885536 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a33c654a-91bc-4019-ac22-b255cb307f9b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.896348 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-scripts" (OuterVolumeSpecName: "scripts") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.896715 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33c654a-91bc-4019-ac22-b255cb307f9b-kube-api-access-5qw2k" (OuterVolumeSpecName: "kube-api-access-5qw2k") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "kube-api-access-5qw2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.918748 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.986523 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qw2k\" (UniqueName: \"kubernetes.io/projected/a33c654a-91bc-4019-ac22-b255cb307f9b-kube-api-access-5qw2k\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.986562 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.986574 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:08 crc kubenswrapper[4979]: I1204 12:06:08.996178 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.022171 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-config-data" (OuterVolumeSpecName: "config-data") pod "a33c654a-91bc-4019-ac22-b255cb307f9b" (UID: "a33c654a-91bc-4019-ac22-b255cb307f9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.088702 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.088746 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33c654a-91bc-4019-ac22-b255cb307f9b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.588669 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a33c654a-91bc-4019-ac22-b255cb307f9b","Type":"ContainerDied","Data":"351e7103733539e18e99b929bb713c1525d4c3b0297bbac75aaa67b4516fad98"} Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.589638 4979 scope.go:117] "RemoveContainer" containerID="f90b541672027c1cbbccae80e9c15acdaa39714638e1bad8203cd07f2d5deaf6" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.588775 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.629008 4979 scope.go:117] "RemoveContainer" containerID="0feb0ba424cc089f915437d27ddadc002d0de23fb58a5ff8475542fdcc9bc7a5" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.629397 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.652476 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.655936 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:09 crc kubenswrapper[4979]: E1204 12:06:09.659274 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-notification-agent" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659309 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-notification-agent" Dec 04 12:06:09 crc kubenswrapper[4979]: E1204 12:06:09.659333 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="proxy-httpd" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659341 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="proxy-httpd" Dec 04 12:06:09 crc kubenswrapper[4979]: E1204 12:06:09.659371 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="sg-core" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659382 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="sg-core" Dec 04 12:06:09 crc kubenswrapper[4979]: E1204 12:06:09.659394 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-central-agent" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659399 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-central-agent" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659558 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="sg-core" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659574 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-central-agent" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659581 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="proxy-httpd" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.659592 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" containerName="ceilometer-notification-agent" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.677392 4979 scope.go:117] "RemoveContainer" containerID="9ba3cd701b799904adceb633b48fe34d2978bdc724cd24f658c7b095673130cb" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.699350 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.704231 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.706220 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.708362 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fklm9\" (UniqueName: \"kubernetes.io/projected/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-kube-api-access-fklm9\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.708499 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-log-httpd\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.708713 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.708750 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.708956 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-config-data\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.709163 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-scripts\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.709206 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-run-httpd\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.714962 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.734912 4979 scope.go:117] "RemoveContainer" containerID="d1787d30e93aaabf4110bd689fcff1607fabe56138f512c93f3a9759e42bb14d" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811071 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811164 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811262 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-config-data\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-scripts\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811366 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-run-httpd\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811412 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fklm9\" (UniqueName: \"kubernetes.io/projected/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-kube-api-access-fklm9\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.811463 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-log-httpd\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.812086 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-log-httpd\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.812133 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-run-httpd\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.816546 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.816640 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.819611 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-config-data\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.821583 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-scripts\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:09 crc kubenswrapper[4979]: I1204 12:06:09.831339 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fklm9\" (UniqueName: \"kubernetes.io/projected/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-kube-api-access-fklm9\") pod \"ceilometer-0\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " pod="openstack/ceilometer-0" Dec 04 12:06:10 crc kubenswrapper[4979]: I1204 12:06:10.024545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:10 crc kubenswrapper[4979]: I1204 12:06:10.209464 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a33c654a-91bc-4019-ac22-b255cb307f9b" path="/var/lib/kubelet/pods/a33c654a-91bc-4019-ac22-b255cb307f9b/volumes" Dec 04 12:06:10 crc kubenswrapper[4979]: I1204 12:06:10.573735 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:10 crc kubenswrapper[4979]: W1204 12:06:10.576906 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd30685c9_3bda_42d2_9a5b_d4ca5a6bcc2b.slice/crio-6e6a7eec92198c13a142cd6e1317018f52a24c211ba7f1be971d545aa8b0923a WatchSource:0}: Error finding container 6e6a7eec92198c13a142cd6e1317018f52a24c211ba7f1be971d545aa8b0923a: Status 404 returned error can't find the container with id 6e6a7eec92198c13a142cd6e1317018f52a24c211ba7f1be971d545aa8b0923a Dec 04 12:06:10 crc kubenswrapper[4979]: I1204 12:06:10.598656 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerStarted","Data":"6e6a7eec92198c13a142cd6e1317018f52a24c211ba7f1be971d545aa8b0923a"} Dec 04 12:06:11 crc kubenswrapper[4979]: I1204 12:06:11.607388 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerStarted","Data":"a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697"} Dec 04 12:06:12 crc kubenswrapper[4979]: I1204 12:06:12.619773 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerStarted","Data":"5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5"} Dec 04 12:06:13 crc kubenswrapper[4979]: I1204 12:06:13.634559 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerStarted","Data":"da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997"} Dec 04 12:06:14 crc kubenswrapper[4979]: I1204 12:06:14.646721 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerStarted","Data":"021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d"} Dec 04 12:06:14 crc kubenswrapper[4979]: I1204 12:06:14.647398 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:06:14 crc kubenswrapper[4979]: I1204 12:06:14.677149 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.109067718 podStartE2EDuration="5.677125251s" podCreationTimestamp="2025-12-04 12:06:09 +0000 UTC" firstStartedPulling="2025-12-04 12:06:10.579000826 +0000 UTC m=+1394.853296630" lastFinishedPulling="2025-12-04 12:06:14.147058359 +0000 UTC m=+1398.421354163" observedRunningTime="2025-12-04 12:06:14.67012374 +0000 UTC m=+1398.944419544" watchObservedRunningTime="2025-12-04 12:06:14.677125251 +0000 UTC m=+1398.951421055" Dec 04 12:06:18 crc kubenswrapper[4979]: I1204 12:06:18.704985 4979 generic.go:334] "Generic (PLEG): container finished" podID="7b62e2c4-065e-48b2-abc0-03f445f8d787" containerID="d6995eed3a9022adb589d900dcb50f240dcc437b7005c2eca54fd004c5fbffa7" exitCode=0 Dec 04 12:06:18 crc kubenswrapper[4979]: I1204 12:06:18.705080 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k59t5" event={"ID":"7b62e2c4-065e-48b2-abc0-03f445f8d787","Type":"ContainerDied","Data":"d6995eed3a9022adb589d900dcb50f240dcc437b7005c2eca54fd004c5fbffa7"} Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.116372 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.248715 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv5pl\" (UniqueName: \"kubernetes.io/projected/7b62e2c4-065e-48b2-abc0-03f445f8d787-kube-api-access-fv5pl\") pod \"7b62e2c4-065e-48b2-abc0-03f445f8d787\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.249505 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-config-data\") pod \"7b62e2c4-065e-48b2-abc0-03f445f8d787\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.249552 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-combined-ca-bundle\") pod \"7b62e2c4-065e-48b2-abc0-03f445f8d787\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.249628 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-scripts\") pod \"7b62e2c4-065e-48b2-abc0-03f445f8d787\" (UID: \"7b62e2c4-065e-48b2-abc0-03f445f8d787\") " Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.254579 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b62e2c4-065e-48b2-abc0-03f445f8d787-kube-api-access-fv5pl" (OuterVolumeSpecName: "kube-api-access-fv5pl") pod "7b62e2c4-065e-48b2-abc0-03f445f8d787" (UID: "7b62e2c4-065e-48b2-abc0-03f445f8d787"). InnerVolumeSpecName "kube-api-access-fv5pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.255159 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-scripts" (OuterVolumeSpecName: "scripts") pod "7b62e2c4-065e-48b2-abc0-03f445f8d787" (UID: "7b62e2c4-065e-48b2-abc0-03f445f8d787"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.278389 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b62e2c4-065e-48b2-abc0-03f445f8d787" (UID: "7b62e2c4-065e-48b2-abc0-03f445f8d787"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.279433 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-config-data" (OuterVolumeSpecName: "config-data") pod "7b62e2c4-065e-48b2-abc0-03f445f8d787" (UID: "7b62e2c4-065e-48b2-abc0-03f445f8d787"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.351341 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.351386 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.351399 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7b62e2c4-065e-48b2-abc0-03f445f8d787-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.351410 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv5pl\" (UniqueName: \"kubernetes.io/projected/7b62e2c4-065e-48b2-abc0-03f445f8d787-kube-api-access-fv5pl\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.727145 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-k59t5" event={"ID":"7b62e2c4-065e-48b2-abc0-03f445f8d787","Type":"ContainerDied","Data":"efec5541d7ec816096769b36b43f21ef73ab6747bfc2c9b94638432440942b93"} Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.727188 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efec5541d7ec816096769b36b43f21ef73ab6747bfc2c9b94638432440942b93" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.727284 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-k59t5" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.838919 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 12:06:20 crc kubenswrapper[4979]: E1204 12:06:20.839372 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b62e2c4-065e-48b2-abc0-03f445f8d787" containerName="nova-cell0-conductor-db-sync" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.839394 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b62e2c4-065e-48b2-abc0-03f445f8d787" containerName="nova-cell0-conductor-db-sync" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.839631 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b62e2c4-065e-48b2-abc0-03f445f8d787" containerName="nova-cell0-conductor-db-sync" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.840353 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.843094 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-vh9gl" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.844784 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.858994 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnlch\" (UniqueName: \"kubernetes.io/projected/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-kube-api-access-hnlch\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.859281 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.859404 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.865267 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.960512 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnlch\" (UniqueName: \"kubernetes.io/projected/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-kube-api-access-hnlch\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.960631 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.960671 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.965717 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.966384 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:20 crc kubenswrapper[4979]: I1204 12:06:20.977632 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnlch\" (UniqueName: \"kubernetes.io/projected/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-kube-api-access-hnlch\") pod \"nova-cell0-conductor-0\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:21 crc kubenswrapper[4979]: I1204 12:06:21.162452 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:21 crc kubenswrapper[4979]: I1204 12:06:21.630232 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 12:06:21 crc kubenswrapper[4979]: I1204 12:06:21.739673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"122bb56e-b8fc-4822-b6f5-6c91122ff4ee","Type":"ContainerStarted","Data":"af9f0679b273274bb1a02d208857e1b9f9cdec84ed8e963065c372d42f9bc905"} Dec 04 12:06:22 crc kubenswrapper[4979]: I1204 12:06:22.747774 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"122bb56e-b8fc-4822-b6f5-6c91122ff4ee","Type":"ContainerStarted","Data":"fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca"} Dec 04 12:06:22 crc kubenswrapper[4979]: I1204 12:06:22.748082 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:22 crc kubenswrapper[4979]: I1204 12:06:22.772255 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.772232421 podStartE2EDuration="2.772232421s" podCreationTimestamp="2025-12-04 12:06:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:22.762513117 +0000 UTC m=+1407.036808911" watchObservedRunningTime="2025-12-04 12:06:22.772232421 +0000 UTC m=+1407.046528235" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.189803 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.630161 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-lcmq5"] Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.631587 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.635630 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.635983 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.648363 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lcmq5"] Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.776134 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbms\" (UniqueName: \"kubernetes.io/projected/814e268c-44eb-4a89-a58f-a38ac92fde57-kube-api-access-kqbms\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.776219 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.776255 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-scripts\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.776327 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-config-data\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.782813 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.784645 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.789617 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.852780 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.878822 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbms\" (UniqueName: \"kubernetes.io/projected/814e268c-44eb-4a89-a58f-a38ac92fde57-kube-api-access-kqbms\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.878897 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.878940 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-scripts\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.878999 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-config-data\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.894379 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-scripts\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.900151 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-config-data\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.903998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.919016 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbms\" (UniqueName: \"kubernetes.io/projected/814e268c-44eb-4a89-a58f-a38ac92fde57-kube-api-access-kqbms\") pod \"nova-cell0-cell-mapping-lcmq5\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.972751 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.982288 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-config-data\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.982602 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.982620 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dbtj\" (UniqueName: \"kubernetes.io/projected/d7167421-ea96-46d2-882d-a4cabc2e4909-kube-api-access-2dbtj\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:26 crc kubenswrapper[4979]: I1204 12:06:26.982664 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7167421-ea96-46d2-882d-a4cabc2e4909-logs\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.083858 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-config-data\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.083940 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dbtj\" (UniqueName: \"kubernetes.io/projected/d7167421-ea96-46d2-882d-a4cabc2e4909-kube-api-access-2dbtj\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.083956 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.084000 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7167421-ea96-46d2-882d-a4cabc2e4909-logs\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.084534 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7167421-ea96-46d2-882d-a4cabc2e4909-logs\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.094787 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-config-data\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.097480 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.130907 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dbtj\" (UniqueName: \"kubernetes.io/projected/d7167421-ea96-46d2-882d-a4cabc2e4909-kube-api-access-2dbtj\") pod \"nova-api-0\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.155794 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.157189 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.162232 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.197398 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.198936 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.200979 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.254405 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.255922 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.258577 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.289435 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-config-data\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.289796 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlbfk\" (UniqueName: \"kubernetes.io/projected/ccbbadd9-c949-43aa-8d63-af3868304f5d-kube-api-access-qlbfk\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.289878 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.297588 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391222 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g48w\" (UniqueName: \"kubernetes.io/projected/fe7b7173-9589-4cc9-94fb-31c77823e6d3-kube-api-access-5g48w\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391270 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391318 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391352 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-config-data\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391413 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-config-data\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391438 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkcxp\" (UniqueName: \"kubernetes.io/projected/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-kube-api-access-dkcxp\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391458 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe7b7173-9589-4cc9-94fb-31c77823e6d3-logs\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391502 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlbfk\" (UniqueName: \"kubernetes.io/projected/ccbbadd9-c949-43aa-8d63-af3868304f5d-kube-api-access-qlbfk\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391551 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391583 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.391607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.400026 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.412039 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-config-data\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.413506 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.424825 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.430919 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlbfk\" (UniqueName: \"kubernetes.io/projected/ccbbadd9-c949-43aa-8d63-af3868304f5d-kube-api-access-qlbfk\") pod \"nova-scheduler-0\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.488853 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.493459 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.493671 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.493895 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g48w\" (UniqueName: \"kubernetes.io/projected/fe7b7173-9589-4cc9-94fb-31c77823e6d3-kube-api-access-5g48w\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.494068 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.494379 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-config-data\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.494500 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkcxp\" (UniqueName: \"kubernetes.io/projected/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-kube-api-access-dkcxp\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.494530 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe7b7173-9589-4cc9-94fb-31c77823e6d3-logs\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.494958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe7b7173-9589-4cc9-94fb-31c77823e6d3-logs\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.497724 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.497993 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.498804 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.506651 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-config-data\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.513895 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q5shf"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.513898 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g48w\" (UniqueName: \"kubernetes.io/projected/fe7b7173-9589-4cc9-94fb-31c77823e6d3-kube-api-access-5g48w\") pod \"nova-metadata-0\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.515425 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.518001 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkcxp\" (UniqueName: \"kubernetes.io/projected/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-kube-api-access-dkcxp\") pod \"nova-cell1-novncproxy-0\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.531213 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q5shf"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.553907 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.634564 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.696926 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-lcmq5"] Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.700812 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-svc\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.700872 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-config\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.700896 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.700978 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5c2k\" (UniqueName: \"kubernetes.io/projected/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-kube-api-access-d5c2k\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.701036 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.701083 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.803096 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-svc\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.803134 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-config\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.803151 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.803205 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5c2k\" (UniqueName: \"kubernetes.io/projected/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-kube-api-access-d5c2k\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.803241 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.803272 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.804089 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-config\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.804187 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-nb\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.804395 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-sb\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.804474 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-swift-storage-0\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.804582 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-svc\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.816074 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lcmq5" event={"ID":"814e268c-44eb-4a89-a58f-a38ac92fde57","Type":"ContainerStarted","Data":"a4e940f0bfe6305b085a73661cf85f709e2238b89f20e86b7ee2b00dfb8a3824"} Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.828334 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5c2k\" (UniqueName: \"kubernetes.io/projected/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-kube-api-access-d5c2k\") pod \"dnsmasq-dns-757b4f8459-q5shf\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:27 crc kubenswrapper[4979]: I1204 12:06:27.855255 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.030794 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:28 crc kubenswrapper[4979]: W1204 12:06:28.034823 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7167421_ea96_46d2_882d_a4cabc2e4909.slice/crio-1d7cfd21c6180eb5b4b97d9d1112cc76a32d9463e94463e379080f796818511e WatchSource:0}: Error finding container 1d7cfd21c6180eb5b4b97d9d1112cc76a32d9463e94463e379080f796818511e: Status 404 returned error can't find the container with id 1d7cfd21c6180eb5b4b97d9d1112cc76a32d9463e94463e379080f796818511e Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.177874 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-26lwp"] Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.179104 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.182629 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.182787 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.191254 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-26lwp"] Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.249900 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.319205 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-scripts\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.319684 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.319804 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlnps\" (UniqueName: \"kubernetes.io/projected/36c9067d-e2c5-443d-922a-78441d66f69b-kube-api-access-nlnps\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.319883 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-config-data\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.372819 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:28 crc kubenswrapper[4979]: W1204 12:06:28.378605 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podccbbadd9_c949_43aa_8d63_af3868304f5d.slice/crio-78dc60a4fb3d8c231be6536fb69331ebc8d8b2f459882f6ef63cc69939b12033 WatchSource:0}: Error finding container 78dc60a4fb3d8c231be6536fb69331ebc8d8b2f459882f6ef63cc69939b12033: Status 404 returned error can't find the container with id 78dc60a4fb3d8c231be6536fb69331ebc8d8b2f459882f6ef63cc69939b12033 Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.388862 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.423407 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.423529 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlnps\" (UniqueName: \"kubernetes.io/projected/36c9067d-e2c5-443d-922a-78441d66f69b-kube-api-access-nlnps\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.423566 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-config-data\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.423636 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-scripts\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.430168 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.431999 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-scripts\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.434775 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-config-data\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.441774 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlnps\" (UniqueName: \"kubernetes.io/projected/36c9067d-e2c5-443d-922a-78441d66f69b-kube-api-access-nlnps\") pod \"nova-cell1-conductor-db-sync-26lwp\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.488615 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q5shf"] Dec 04 12:06:28 crc kubenswrapper[4979]: W1204 12:06:28.495339 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb31b83e1_6787_4470_94f7_dcdfd9dad5aa.slice/crio-1abf45c809845bce062ae2dda266dc68c49e8502eaf5a2256db17f698fbff73a WatchSource:0}: Error finding container 1abf45c809845bce062ae2dda266dc68c49e8502eaf5a2256db17f698fbff73a: Status 404 returned error can't find the container with id 1abf45c809845bce062ae2dda266dc68c49e8502eaf5a2256db17f698fbff73a Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.501174 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.826529 4979 generic.go:334] "Generic (PLEG): container finished" podID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerID="68bb06f1ab863da3c41eee87d1183cafa70bd684d96aec88da53971c3cdc2983" exitCode=0 Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.826922 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" event={"ID":"b31b83e1-6787-4470-94f7-dcdfd9dad5aa","Type":"ContainerDied","Data":"68bb06f1ab863da3c41eee87d1183cafa70bd684d96aec88da53971c3cdc2983"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.828501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" event={"ID":"b31b83e1-6787-4470-94f7-dcdfd9dad5aa","Type":"ContainerStarted","Data":"1abf45c809845bce062ae2dda266dc68c49e8502eaf5a2256db17f698fbff73a"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.834678 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7167421-ea96-46d2-882d-a4cabc2e4909","Type":"ContainerStarted","Data":"1d7cfd21c6180eb5b4b97d9d1112cc76a32d9463e94463e379080f796818511e"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.836258 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ccbbadd9-c949-43aa-8d63-af3868304f5d","Type":"ContainerStarted","Data":"78dc60a4fb3d8c231be6536fb69331ebc8d8b2f459882f6ef63cc69939b12033"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.838658 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lcmq5" event={"ID":"814e268c-44eb-4a89-a58f-a38ac92fde57","Type":"ContainerStarted","Data":"6466c98b4d2b75be62b2edb44c9e2313bed08ae164cacc3957a45c111ebc7cc6"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.842425 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af","Type":"ContainerStarted","Data":"453b25852a4752b93720b184ea15bea72967462fa507404d355fc30313bfd88f"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.843843 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe7b7173-9589-4cc9-94fb-31c77823e6d3","Type":"ContainerStarted","Data":"1455faa21fc6e4c571f376da2d2361d7ae385552eb4c0cc5e9188797ec5aea88"} Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.881343 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-lcmq5" podStartSLOduration=2.881325783 podStartE2EDuration="2.881325783s" podCreationTimestamp="2025-12-04 12:06:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:28.874466986 +0000 UTC m=+1413.148762790" watchObservedRunningTime="2025-12-04 12:06:28.881325783 +0000 UTC m=+1413.155621587" Dec 04 12:06:28 crc kubenswrapper[4979]: I1204 12:06:28.990204 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-26lwp"] Dec 04 12:06:29 crc kubenswrapper[4979]: W1204 12:06:29.005066 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36c9067d_e2c5_443d_922a_78441d66f69b.slice/crio-17327e31253b156d79bb4ab0e70a6792b4d9a862958edfd5ac9b10ba69a60c58 WatchSource:0}: Error finding container 17327e31253b156d79bb4ab0e70a6792b4d9a862958edfd5ac9b10ba69a60c58: Status 404 returned error can't find the container with id 17327e31253b156d79bb4ab0e70a6792b4d9a862958edfd5ac9b10ba69a60c58 Dec 04 12:06:29 crc kubenswrapper[4979]: I1204 12:06:29.855775 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-26lwp" event={"ID":"36c9067d-e2c5-443d-922a-78441d66f69b","Type":"ContainerStarted","Data":"716bf63a8969da92127307155cf7ce295dd76465e78ce82e446d08fcdb27f11a"} Dec 04 12:06:29 crc kubenswrapper[4979]: I1204 12:06:29.856151 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-26lwp" event={"ID":"36c9067d-e2c5-443d-922a-78441d66f69b","Type":"ContainerStarted","Data":"17327e31253b156d79bb4ab0e70a6792b4d9a862958edfd5ac9b10ba69a60c58"} Dec 04 12:06:29 crc kubenswrapper[4979]: I1204 12:06:29.861154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" event={"ID":"b31b83e1-6787-4470-94f7-dcdfd9dad5aa","Type":"ContainerStarted","Data":"f17eef864c09a58ad72a34fd9c6c3191ef45028bc4ee2edec5f5297016f9e2a8"} Dec 04 12:06:29 crc kubenswrapper[4979]: I1204 12:06:29.861202 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:29 crc kubenswrapper[4979]: I1204 12:06:29.875911 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-26lwp" podStartSLOduration=1.8758946440000002 podStartE2EDuration="1.875894644s" podCreationTimestamp="2025-12-04 12:06:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:29.873617223 +0000 UTC m=+1414.147913027" watchObservedRunningTime="2025-12-04 12:06:29.875894644 +0000 UTC m=+1414.150190448" Dec 04 12:06:29 crc kubenswrapper[4979]: I1204 12:06:29.929469 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" podStartSLOduration=2.929444471 podStartE2EDuration="2.929444471s" podCreationTimestamp="2025-12-04 12:06:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:29.916331134 +0000 UTC m=+1414.190626958" watchObservedRunningTime="2025-12-04 12:06:29.929444471 +0000 UTC m=+1414.203740275" Dec 04 12:06:30 crc kubenswrapper[4979]: I1204 12:06:30.455831 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:06:30 crc kubenswrapper[4979]: I1204 12:06:30.463944 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.891895 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe7b7173-9589-4cc9-94fb-31c77823e6d3","Type":"ContainerStarted","Data":"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d"} Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.892440 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe7b7173-9589-4cc9-94fb-31c77823e6d3","Type":"ContainerStarted","Data":"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860"} Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.892063 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-metadata" containerID="cri-o://693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d" gracePeriod=30 Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.891974 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-log" containerID="cri-o://385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860" gracePeriod=30 Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.895956 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7167421-ea96-46d2-882d-a4cabc2e4909","Type":"ContainerStarted","Data":"93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744"} Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.896329 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7167421-ea96-46d2-882d-a4cabc2e4909","Type":"ContainerStarted","Data":"25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae"} Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.899073 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ccbbadd9-c949-43aa-8d63-af3868304f5d","Type":"ContainerStarted","Data":"d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158"} Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.901087 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af","Type":"ContainerStarted","Data":"ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8"} Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.901211 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8" gracePeriod=30 Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.920788 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.622190856 podStartE2EDuration="5.920761642s" podCreationTimestamp="2025-12-04 12:06:27 +0000 UTC" firstStartedPulling="2025-12-04 12:06:28.389404268 +0000 UTC m=+1412.663700072" lastFinishedPulling="2025-12-04 12:06:31.687975054 +0000 UTC m=+1415.962270858" observedRunningTime="2025-12-04 12:06:32.915435927 +0000 UTC m=+1417.189731731" watchObservedRunningTime="2025-12-04 12:06:32.920761642 +0000 UTC m=+1417.195057456" Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.946383 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.294938588 podStartE2EDuration="6.946362948s" podCreationTimestamp="2025-12-04 12:06:26 +0000 UTC" firstStartedPulling="2025-12-04 12:06:28.037487829 +0000 UTC m=+1412.311783633" lastFinishedPulling="2025-12-04 12:06:31.688912169 +0000 UTC m=+1415.963207993" observedRunningTime="2025-12-04 12:06:32.940116359 +0000 UTC m=+1417.214412173" watchObservedRunningTime="2025-12-04 12:06:32.946362948 +0000 UTC m=+1417.220658752" Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.962103 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.653716893 podStartE2EDuration="5.962085636s" podCreationTimestamp="2025-12-04 12:06:27 +0000 UTC" firstStartedPulling="2025-12-04 12:06:28.380004462 +0000 UTC m=+1412.654300276" lastFinishedPulling="2025-12-04 12:06:31.688373215 +0000 UTC m=+1415.962669019" observedRunningTime="2025-12-04 12:06:32.956277638 +0000 UTC m=+1417.230573442" watchObservedRunningTime="2025-12-04 12:06:32.962085636 +0000 UTC m=+1417.236381440" Dec 04 12:06:32 crc kubenswrapper[4979]: I1204 12:06:32.980217 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.54622442 podStartE2EDuration="5.980200048s" podCreationTimestamp="2025-12-04 12:06:27 +0000 UTC" firstStartedPulling="2025-12-04 12:06:28.259590138 +0000 UTC m=+1412.533885932" lastFinishedPulling="2025-12-04 12:06:31.693565756 +0000 UTC m=+1415.967861560" observedRunningTime="2025-12-04 12:06:32.972555391 +0000 UTC m=+1417.246851195" watchObservedRunningTime="2025-12-04 12:06:32.980200048 +0000 UTC m=+1417.254495852" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.470363 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.560994 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe7b7173-9589-4cc9-94fb-31c77823e6d3-logs\") pod \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.561107 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g48w\" (UniqueName: \"kubernetes.io/projected/fe7b7173-9589-4cc9-94fb-31c77823e6d3-kube-api-access-5g48w\") pod \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.561212 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-combined-ca-bundle\") pod \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.561279 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-config-data\") pod \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\" (UID: \"fe7b7173-9589-4cc9-94fb-31c77823e6d3\") " Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.561436 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe7b7173-9589-4cc9-94fb-31c77823e6d3-logs" (OuterVolumeSpecName: "logs") pod "fe7b7173-9589-4cc9-94fb-31c77823e6d3" (UID: "fe7b7173-9589-4cc9-94fb-31c77823e6d3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.561672 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fe7b7173-9589-4cc9-94fb-31c77823e6d3-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.568410 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe7b7173-9589-4cc9-94fb-31c77823e6d3-kube-api-access-5g48w" (OuterVolumeSpecName: "kube-api-access-5g48w") pod "fe7b7173-9589-4cc9-94fb-31c77823e6d3" (UID: "fe7b7173-9589-4cc9-94fb-31c77823e6d3"). InnerVolumeSpecName "kube-api-access-5g48w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.598049 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-config-data" (OuterVolumeSpecName: "config-data") pod "fe7b7173-9589-4cc9-94fb-31c77823e6d3" (UID: "fe7b7173-9589-4cc9-94fb-31c77823e6d3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.603783 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fe7b7173-9589-4cc9-94fb-31c77823e6d3" (UID: "fe7b7173-9589-4cc9-94fb-31c77823e6d3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.663917 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.663959 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g48w\" (UniqueName: \"kubernetes.io/projected/fe7b7173-9589-4cc9-94fb-31c77823e6d3-kube-api-access-5g48w\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.663969 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe7b7173-9589-4cc9-94fb-31c77823e6d3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915535 4979 generic.go:334] "Generic (PLEG): container finished" podID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerID="693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d" exitCode=0 Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915562 4979 generic.go:334] "Generic (PLEG): container finished" podID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerID="385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860" exitCode=143 Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915602 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe7b7173-9589-4cc9-94fb-31c77823e6d3","Type":"ContainerDied","Data":"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d"} Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915680 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe7b7173-9589-4cc9-94fb-31c77823e6d3","Type":"ContainerDied","Data":"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860"} Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915697 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fe7b7173-9589-4cc9-94fb-31c77823e6d3","Type":"ContainerDied","Data":"1455faa21fc6e4c571f376da2d2361d7ae385552eb4c0cc5e9188797ec5aea88"} Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915623 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.915732 4979 scope.go:117] "RemoveContainer" containerID="693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.935411 4979 scope.go:117] "RemoveContainer" containerID="385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.958172 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.966294 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.969896 4979 scope.go:117] "RemoveContainer" containerID="693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d" Dec 04 12:06:33 crc kubenswrapper[4979]: E1204 12:06:33.970324 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d\": container with ID starting with 693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d not found: ID does not exist" containerID="693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.970362 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d"} err="failed to get container status \"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d\": rpc error: code = NotFound desc = could not find container \"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d\": container with ID starting with 693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d not found: ID does not exist" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.970388 4979 scope.go:117] "RemoveContainer" containerID="385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860" Dec 04 12:06:33 crc kubenswrapper[4979]: E1204 12:06:33.970703 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860\": container with ID starting with 385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860 not found: ID does not exist" containerID="385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.970730 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860"} err="failed to get container status \"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860\": rpc error: code = NotFound desc = could not find container \"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860\": container with ID starting with 385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860 not found: ID does not exist" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.970747 4979 scope.go:117] "RemoveContainer" containerID="693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.971076 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d"} err="failed to get container status \"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d\": rpc error: code = NotFound desc = could not find container \"693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d\": container with ID starting with 693a74cdf51adbc405b11ceceda858abd0787b50f253ed04871101c8590a7a5d not found: ID does not exist" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.971101 4979 scope.go:117] "RemoveContainer" containerID="385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.971485 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860"} err="failed to get container status \"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860\": rpc error: code = NotFound desc = could not find container \"385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860\": container with ID starting with 385d84caebc391932da5fc4fe173172b6112c76b987c2a7f8bff14c132d91860 not found: ID does not exist" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.984408 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:33 crc kubenswrapper[4979]: E1204 12:06:33.984861 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-metadata" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.984881 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-metadata" Dec 04 12:06:33 crc kubenswrapper[4979]: E1204 12:06:33.984899 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-log" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.984907 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-log" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.985205 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-log" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.985229 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" containerName="nova-metadata-metadata" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.986197 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.989373 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 12:06:33 crc kubenswrapper[4979]: I1204 12:06:33.989995 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.002213 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.071986 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.072046 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.072139 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-config-data\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.072158 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz97r\" (UniqueName: \"kubernetes.io/projected/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-kube-api-access-hz97r\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.072237 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-logs\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.174213 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.174265 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.174351 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-config-data\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.174375 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz97r\" (UniqueName: \"kubernetes.io/projected/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-kube-api-access-hz97r\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.174456 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-logs\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.174949 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-logs\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.178893 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.179553 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-config-data\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.189894 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.201726 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz97r\" (UniqueName: \"kubernetes.io/projected/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-kube-api-access-hz97r\") pod \"nova-metadata-0\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.217189 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe7b7173-9589-4cc9-94fb-31c77823e6d3" path="/var/lib/kubelet/pods/fe7b7173-9589-4cc9-94fb-31c77823e6d3/volumes" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.300818 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.769005 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:34 crc kubenswrapper[4979]: I1204 12:06:34.927792 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf1ea4ae-4bd1-4898-a6b3-1befe6553086","Type":"ContainerStarted","Data":"87c68ccca481e093c3480d42f8a4d043c448d4f84f1fa4fe1d52a918959b3967"} Dec 04 12:06:35 crc kubenswrapper[4979]: I1204 12:06:35.941437 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf1ea4ae-4bd1-4898-a6b3-1befe6553086","Type":"ContainerStarted","Data":"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a"} Dec 04 12:06:35 crc kubenswrapper[4979]: I1204 12:06:35.941949 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf1ea4ae-4bd1-4898-a6b3-1befe6553086","Type":"ContainerStarted","Data":"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f"} Dec 04 12:06:35 crc kubenswrapper[4979]: I1204 12:06:35.942958 4979 generic.go:334] "Generic (PLEG): container finished" podID="814e268c-44eb-4a89-a58f-a38ac92fde57" containerID="6466c98b4d2b75be62b2edb44c9e2313bed08ae164cacc3957a45c111ebc7cc6" exitCode=0 Dec 04 12:06:35 crc kubenswrapper[4979]: I1204 12:06:35.942987 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lcmq5" event={"ID":"814e268c-44eb-4a89-a58f-a38ac92fde57","Type":"ContainerDied","Data":"6466c98b4d2b75be62b2edb44c9e2313bed08ae164cacc3957a45c111ebc7cc6"} Dec 04 12:06:35 crc kubenswrapper[4979]: I1204 12:06:35.968741 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.9687208739999997 podStartE2EDuration="2.968720874s" podCreationTimestamp="2025-12-04 12:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:35.959796432 +0000 UTC m=+1420.234092236" watchObservedRunningTime="2025-12-04 12:06:35.968720874 +0000 UTC m=+1420.243016688" Dec 04 12:06:36 crc kubenswrapper[4979]: I1204 12:06:36.958694 4979 generic.go:334] "Generic (PLEG): container finished" podID="36c9067d-e2c5-443d-922a-78441d66f69b" containerID="716bf63a8969da92127307155cf7ce295dd76465e78ce82e446d08fcdb27f11a" exitCode=0 Dec 04 12:06:36 crc kubenswrapper[4979]: I1204 12:06:36.958800 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-26lwp" event={"ID":"36c9067d-e2c5-443d-922a-78441d66f69b","Type":"ContainerDied","Data":"716bf63a8969da92127307155cf7ce295dd76465e78ce82e446d08fcdb27f11a"} Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.307131 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.400542 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.400585 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.440970 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-config-data\") pod \"814e268c-44eb-4a89-a58f-a38ac92fde57\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.441090 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-scripts\") pod \"814e268c-44eb-4a89-a58f-a38ac92fde57\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.441164 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-combined-ca-bundle\") pod \"814e268c-44eb-4a89-a58f-a38ac92fde57\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.441226 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqbms\" (UniqueName: \"kubernetes.io/projected/814e268c-44eb-4a89-a58f-a38ac92fde57-kube-api-access-kqbms\") pod \"814e268c-44eb-4a89-a58f-a38ac92fde57\" (UID: \"814e268c-44eb-4a89-a58f-a38ac92fde57\") " Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.447784 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/814e268c-44eb-4a89-a58f-a38ac92fde57-kube-api-access-kqbms" (OuterVolumeSpecName: "kube-api-access-kqbms") pod "814e268c-44eb-4a89-a58f-a38ac92fde57" (UID: "814e268c-44eb-4a89-a58f-a38ac92fde57"). InnerVolumeSpecName "kube-api-access-kqbms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.453472 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-scripts" (OuterVolumeSpecName: "scripts") pod "814e268c-44eb-4a89-a58f-a38ac92fde57" (UID: "814e268c-44eb-4a89-a58f-a38ac92fde57"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.475789 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-config-data" (OuterVolumeSpecName: "config-data") pod "814e268c-44eb-4a89-a58f-a38ac92fde57" (UID: "814e268c-44eb-4a89-a58f-a38ac92fde57"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.479450 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "814e268c-44eb-4a89-a58f-a38ac92fde57" (UID: "814e268c-44eb-4a89-a58f-a38ac92fde57"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.490149 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.491347 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.519817 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.547739 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.547790 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.547803 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqbms\" (UniqueName: \"kubernetes.io/projected/814e268c-44eb-4a89-a58f-a38ac92fde57-kube-api-access-kqbms\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.547818 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/814e268c-44eb-4a89-a58f-a38ac92fde57-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.554723 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.857527 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.945448 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-95sg6"] Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.945708 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerName="dnsmasq-dns" containerID="cri-o://26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6" gracePeriod=10 Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.973269 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-lcmq5" Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.976876 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-lcmq5" event={"ID":"814e268c-44eb-4a89-a58f-a38ac92fde57","Type":"ContainerDied","Data":"a4e940f0bfe6305b085a73661cf85f709e2238b89f20e86b7ee2b00dfb8a3824"} Dec 04 12:06:37 crc kubenswrapper[4979]: I1204 12:06:37.976935 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4e940f0bfe6305b085a73661cf85f709e2238b89f20e86b7ee2b00dfb8a3824" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.017983 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.168751 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.169276 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-log" containerID="cri-o://25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae" gracePeriod=30 Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.169354 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-api" containerID="cri-o://93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744" gracePeriod=30 Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.175316 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.179854 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.183:8774/\": EOF" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.191339 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.191550 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-log" containerID="cri-o://1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f" gracePeriod=30 Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.192019 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-metadata" containerID="cri-o://409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a" gracePeriod=30 Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.607628 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.620142 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.635404 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.677915 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-combined-ca-bundle\") pod \"36c9067d-e2c5-443d-922a-78441d66f69b\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.678001 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-scripts\") pod \"36c9067d-e2c5-443d-922a-78441d66f69b\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.678062 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-config-data\") pod \"36c9067d-e2c5-443d-922a-78441d66f69b\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.678119 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nlnps\" (UniqueName: \"kubernetes.io/projected/36c9067d-e2c5-443d-922a-78441d66f69b-kube-api-access-nlnps\") pod \"36c9067d-e2c5-443d-922a-78441d66f69b\" (UID: \"36c9067d-e2c5-443d-922a-78441d66f69b\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.684135 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36c9067d-e2c5-443d-922a-78441d66f69b-kube-api-access-nlnps" (OuterVolumeSpecName: "kube-api-access-nlnps") pod "36c9067d-e2c5-443d-922a-78441d66f69b" (UID: "36c9067d-e2c5-443d-922a-78441d66f69b"). InnerVolumeSpecName "kube-api-access-nlnps". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.684826 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-scripts" (OuterVolumeSpecName: "scripts") pod "36c9067d-e2c5-443d-922a-78441d66f69b" (UID: "36c9067d-e2c5-443d-922a-78441d66f69b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.726215 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "36c9067d-e2c5-443d-922a-78441d66f69b" (UID: "36c9067d-e2c5-443d-922a-78441d66f69b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.743556 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-config-data" (OuterVolumeSpecName: "config-data") pod "36c9067d-e2c5-443d-922a-78441d66f69b" (UID: "36c9067d-e2c5-443d-922a-78441d66f69b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.774416 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.779178 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48plj\" (UniqueName: \"kubernetes.io/projected/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-kube-api-access-48plj\") pod \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.779327 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-config\") pod \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.779452 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-nb\") pod \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.779482 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-sb\") pod \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.779930 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-svc\") pod \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.779985 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-swift-storage-0\") pod \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\" (UID: \"b4c7c37c-3ab7-43db-ad20-8cd6df70395a\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.780457 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nlnps\" (UniqueName: \"kubernetes.io/projected/36c9067d-e2c5-443d-922a-78441d66f69b-kube-api-access-nlnps\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.780476 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.780487 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.780507 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36c9067d-e2c5-443d-922a-78441d66f69b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.790521 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-kube-api-access-48plj" (OuterVolumeSpecName: "kube-api-access-48plj") pod "b4c7c37c-3ab7-43db-ad20-8cd6df70395a" (UID: "b4c7c37c-3ab7-43db-ad20-8cd6df70395a"). InnerVolumeSpecName "kube-api-access-48plj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.851147 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4c7c37c-3ab7-43db-ad20-8cd6df70395a" (UID: "b4c7c37c-3ab7-43db-ad20-8cd6df70395a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.851449 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-config" (OuterVolumeSpecName: "config") pod "b4c7c37c-3ab7-43db-ad20-8cd6df70395a" (UID: "b4c7c37c-3ab7-43db-ad20-8cd6df70395a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.854095 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4c7c37c-3ab7-43db-ad20-8cd6df70395a" (UID: "b4c7c37c-3ab7-43db-ad20-8cd6df70395a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.854826 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4c7c37c-3ab7-43db-ad20-8cd6df70395a" (UID: "b4c7c37c-3ab7-43db-ad20-8cd6df70395a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.864064 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b4c7c37c-3ab7-43db-ad20-8cd6df70395a" (UID: "b4c7c37c-3ab7-43db-ad20-8cd6df70395a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.894207 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz97r\" (UniqueName: \"kubernetes.io/projected/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-kube-api-access-hz97r\") pod \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.894348 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-config-data\") pod \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.894387 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-nova-metadata-tls-certs\") pod \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.894456 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-logs\") pod \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.894645 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-combined-ca-bundle\") pod \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\" (UID: \"bf1ea4ae-4bd1-4898-a6b3-1befe6553086\") " Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.894810 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-logs" (OuterVolumeSpecName: "logs") pod "bf1ea4ae-4bd1-4898-a6b3-1befe6553086" (UID: "bf1ea4ae-4bd1-4898-a6b3-1befe6553086"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895166 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895189 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895201 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895213 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895226 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48plj\" (UniqueName: \"kubernetes.io/projected/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-kube-api-access-48plj\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895238 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.895249 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4c7c37c-3ab7-43db-ad20-8cd6df70395a-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.901512 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-kube-api-access-hz97r" (OuterVolumeSpecName: "kube-api-access-hz97r") pod "bf1ea4ae-4bd1-4898-a6b3-1befe6553086" (UID: "bf1ea4ae-4bd1-4898-a6b3-1befe6553086"). InnerVolumeSpecName "kube-api-access-hz97r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.967481 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf1ea4ae-4bd1-4898-a6b3-1befe6553086" (UID: "bf1ea4ae-4bd1-4898-a6b3-1befe6553086"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.976316 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-config-data" (OuterVolumeSpecName: "config-data") pod "bf1ea4ae-4bd1-4898-a6b3-1befe6553086" (UID: "bf1ea4ae-4bd1-4898-a6b3-1befe6553086"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.996836 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.996867 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:38 crc kubenswrapper[4979]: I1204 12:06:38.996876 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz97r\" (UniqueName: \"kubernetes.io/projected/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-kube-api-access-hz97r\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.008346 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-26lwp" event={"ID":"36c9067d-e2c5-443d-922a-78441d66f69b","Type":"ContainerDied","Data":"17327e31253b156d79bb4ab0e70a6792b4d9a862958edfd5ac9b10ba69a60c58"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.008398 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17327e31253b156d79bb4ab0e70a6792b4d9a862958edfd5ac9b10ba69a60c58" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.008487 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-26lwp" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.008477 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "bf1ea4ae-4bd1-4898-a6b3-1befe6553086" (UID: "bf1ea4ae-4bd1-4898-a6b3-1befe6553086"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.039599 4979 generic.go:334] "Generic (PLEG): container finished" podID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerID="26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6" exitCode=0 Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.039696 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" event={"ID":"b4c7c37c-3ab7-43db-ad20-8cd6df70395a","Type":"ContainerDied","Data":"26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.039730 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" event={"ID":"b4c7c37c-3ab7-43db-ad20-8cd6df70395a","Type":"ContainerDied","Data":"80f5dd7780f6820011c49d7d051b71260ba35d77206cc9882d5c9c7961c8a0fd"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.039752 4979 scope.go:117] "RemoveContainer" containerID="26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.039915 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-95sg6" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.082910 4979 generic.go:334] "Generic (PLEG): container finished" podID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerID="25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae" exitCode=143 Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.083019 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7167421-ea96-46d2-882d-a4cabc2e4909","Type":"ContainerDied","Data":"25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.095613 4979 scope.go:117] "RemoveContainer" containerID="2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.098907 4979 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf1ea4ae-4bd1-4898-a6b3-1befe6553086-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.101456 4979 generic.go:334] "Generic (PLEG): container finished" podID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerID="409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a" exitCode=0 Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.101486 4979 generic.go:334] "Generic (PLEG): container finished" podID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerID="1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f" exitCode=143 Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.102688 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.104412 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf1ea4ae-4bd1-4898-a6b3-1befe6553086","Type":"ContainerDied","Data":"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.104503 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf1ea4ae-4bd1-4898-a6b3-1befe6553086","Type":"ContainerDied","Data":"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.104521 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bf1ea4ae-4bd1-4898-a6b3-1befe6553086","Type":"ContainerDied","Data":"87c68ccca481e093c3480d42f8a4d043c448d4f84f1fa4fe1d52a918959b3967"} Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.110006 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-95sg6"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.121720 4979 scope.go:117] "RemoveContainer" containerID="26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.125754 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6\": container with ID starting with 26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6 not found: ID does not exist" containerID="26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.125816 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6"} err="failed to get container status \"26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6\": rpc error: code = NotFound desc = could not find container \"26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6\": container with ID starting with 26876a06f1a3bc25480662ad61d15703f2701c73781e57c64a37d1f2f3c190a6 not found: ID does not exist" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.125849 4979 scope.go:117] "RemoveContainer" containerID="2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.130418 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a\": container with ID starting with 2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a not found: ID does not exist" containerID="2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.130465 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a"} err="failed to get container status \"2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a\": rpc error: code = NotFound desc = could not find container \"2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a\": container with ID starting with 2e2191443f638bf7c7d163eb8fbdf933c8f74c819c91b32e74e9df1bffc46b3a not found: ID does not exist" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.130495 4979 scope.go:117] "RemoveContainer" containerID="409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.139084 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-95sg6"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.169035 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.181704 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerName="init" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.181897 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerName="init" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.181984 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerName="dnsmasq-dns" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182035 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerName="dnsmasq-dns" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.182103 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-log" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182171 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-log" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.182234 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="814e268c-44eb-4a89-a58f-a38ac92fde57" containerName="nova-manage" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182293 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="814e268c-44eb-4a89-a58f-a38ac92fde57" containerName="nova-manage" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.182382 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36c9067d-e2c5-443d-922a-78441d66f69b" containerName="nova-cell1-conductor-db-sync" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182441 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="36c9067d-e2c5-443d-922a-78441d66f69b" containerName="nova-cell1-conductor-db-sync" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.182507 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-metadata" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182563 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-metadata" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182816 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="814e268c-44eb-4a89-a58f-a38ac92fde57" containerName="nova-manage" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182881 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-metadata" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.182944 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" containerName="dnsmasq-dns" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.183013 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" containerName="nova-metadata-log" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.183073 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="36c9067d-e2c5-443d-922a-78441d66f69b" containerName="nova-cell1-conductor-db-sync" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.183838 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.193750 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.199033 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.216319 4979 scope.go:117] "RemoveContainer" containerID="1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.220100 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.228185 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.237160 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.238928 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.241263 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.241583 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.250504 4979 scope.go:117] "RemoveContainer" containerID="409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.250861 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a\": container with ID starting with 409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a not found: ID does not exist" containerID="409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.250966 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a"} err="failed to get container status \"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a\": rpc error: code = NotFound desc = could not find container \"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a\": container with ID starting with 409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a not found: ID does not exist" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.251054 4979 scope.go:117] "RemoveContainer" containerID="1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f" Dec 04 12:06:39 crc kubenswrapper[4979]: E1204 12:06:39.251329 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f\": container with ID starting with 1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f not found: ID does not exist" containerID="1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.251427 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f"} err="failed to get container status \"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f\": rpc error: code = NotFound desc = could not find container \"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f\": container with ID starting with 1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f not found: ID does not exist" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.251491 4979 scope.go:117] "RemoveContainer" containerID="409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.251787 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a"} err="failed to get container status \"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a\": rpc error: code = NotFound desc = could not find container \"409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a\": container with ID starting with 409deb2c019ddcad40fc67277641e362f87b240f0bd413f1c787dcf8bca5906a not found: ID does not exist" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.251831 4979 scope.go:117] "RemoveContainer" containerID="1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.252080 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f"} err="failed to get container status \"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f\": rpc error: code = NotFound desc = could not find container \"1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f\": container with ID starting with 1d0a20a8f10a8202e97b4fc5b94f471867bb33e691a4d0e04a44b5666817b36f not found: ID does not exist" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.252677 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306369 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zwdr\" (UniqueName: \"kubernetes.io/projected/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-kube-api-access-7zwdr\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306442 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptrdf\" (UniqueName: \"kubernetes.io/projected/bf314b53-497e-47cd-adce-c905fff3a5b7-kube-api-access-ptrdf\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306491 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-config-data\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306507 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306556 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-logs\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306618 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306642 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.306685 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408476 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zwdr\" (UniqueName: \"kubernetes.io/projected/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-kube-api-access-7zwdr\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptrdf\" (UniqueName: \"kubernetes.io/projected/bf314b53-497e-47cd-adce-c905fff3a5b7-kube-api-access-ptrdf\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408597 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-config-data\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408623 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408682 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-logs\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408750 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408778 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.408820 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.409290 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-logs\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.413039 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.413081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-config-data\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.413409 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.427178 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.427178 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.430275 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zwdr\" (UniqueName: \"kubernetes.io/projected/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-kube-api-access-7zwdr\") pod \"nova-metadata-0\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " pod="openstack/nova-metadata-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.430451 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptrdf\" (UniqueName: \"kubernetes.io/projected/bf314b53-497e-47cd-adce-c905fff3a5b7-kube-api-access-ptrdf\") pod \"nova-cell1-conductor-0\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.522475 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:39 crc kubenswrapper[4979]: I1204 12:06:39.567930 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.033341 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.039411 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 12:06:40 crc kubenswrapper[4979]: W1204 12:06:40.042388 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf314b53_497e_47cd_adce_c905fff3a5b7.slice/crio-5b3488c9494646b6d351ac4a684d1daf4d8f4e67ae844650672f0bc8f67b7be9 WatchSource:0}: Error finding container 5b3488c9494646b6d351ac4a684d1daf4d8f4e67ae844650672f0bc8f67b7be9: Status 404 returned error can't find the container with id 5b3488c9494646b6d351ac4a684d1daf4d8f4e67ae844650672f0bc8f67b7be9 Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.133744 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.133927 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bf314b53-497e-47cd-adce-c905fff3a5b7","Type":"ContainerStarted","Data":"5b3488c9494646b6d351ac4a684d1daf4d8f4e67ae844650672f0bc8f67b7be9"} Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.137391 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ccbbadd9-c949-43aa-8d63-af3868304f5d" containerName="nova-scheduler-scheduler" containerID="cri-o://d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" gracePeriod=30 Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.217097 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4c7c37c-3ab7-43db-ad20-8cd6df70395a" path="/var/lib/kubelet/pods/b4c7c37c-3ab7-43db-ad20-8cd6df70395a/volumes" Dec 04 12:06:40 crc kubenswrapper[4979]: I1204 12:06:40.217721 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf1ea4ae-4bd1-4898-a6b3-1befe6553086" path="/var/lib/kubelet/pods/bf1ea4ae-4bd1-4898-a6b3-1befe6553086/volumes" Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.147642 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab","Type":"ContainerStarted","Data":"24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be"} Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.147973 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab","Type":"ContainerStarted","Data":"ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba"} Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.147983 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab","Type":"ContainerStarted","Data":"da3e0d2b0e1bdedade32c382dfefbbd15d605da522f9f4b7c2c020f985400ecd"} Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.149979 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bf314b53-497e-47cd-adce-c905fff3a5b7","Type":"ContainerStarted","Data":"8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d"} Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.150119 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.167520 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.167504746 podStartE2EDuration="2.167504746s" podCreationTimestamp="2025-12-04 12:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:41.164622947 +0000 UTC m=+1425.438918751" watchObservedRunningTime="2025-12-04 12:06:41.167504746 +0000 UTC m=+1425.441800550" Dec 04 12:06:41 crc kubenswrapper[4979]: I1204 12:06:41.182735 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.182713629 podStartE2EDuration="2.182713629s" podCreationTimestamp="2025-12-04 12:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:41.181981379 +0000 UTC m=+1425.456277193" watchObservedRunningTime="2025-12-04 12:06:41.182713629 +0000 UTC m=+1425.457009443" Dec 04 12:06:42 crc kubenswrapper[4979]: E1204 12:06:42.490985 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:06:42 crc kubenswrapper[4979]: E1204 12:06:42.493875 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:06:42 crc kubenswrapper[4979]: E1204 12:06:42.495157 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:06:42 crc kubenswrapper[4979]: E1204 12:06:42.495221 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ccbbadd9-c949-43aa-8d63-af3868304f5d" containerName="nova-scheduler-scheduler" Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.556016 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.556569 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="dd8830de-44cb-4015-98f2-2d55ffdccf52" containerName="kube-state-metrics" containerID="cri-o://54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036" gracePeriod=30 Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.758179 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.850642 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-combined-ca-bundle\") pod \"ccbbadd9-c949-43aa-8d63-af3868304f5d\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.850733 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-config-data\") pod \"ccbbadd9-c949-43aa-8d63-af3868304f5d\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.850809 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlbfk\" (UniqueName: \"kubernetes.io/projected/ccbbadd9-c949-43aa-8d63-af3868304f5d-kube-api-access-qlbfk\") pod \"ccbbadd9-c949-43aa-8d63-af3868304f5d\" (UID: \"ccbbadd9-c949-43aa-8d63-af3868304f5d\") " Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.880888 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccbbadd9-c949-43aa-8d63-af3868304f5d-kube-api-access-qlbfk" (OuterVolumeSpecName: "kube-api-access-qlbfk") pod "ccbbadd9-c949-43aa-8d63-af3868304f5d" (UID: "ccbbadd9-c949-43aa-8d63-af3868304f5d"). InnerVolumeSpecName "kube-api-access-qlbfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.961589 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlbfk\" (UniqueName: \"kubernetes.io/projected/ccbbadd9-c949-43aa-8d63-af3868304f5d-kube-api-access-qlbfk\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:43 crc kubenswrapper[4979]: I1204 12:06:43.989944 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-config-data" (OuterVolumeSpecName: "config-data") pod "ccbbadd9-c949-43aa-8d63-af3868304f5d" (UID: "ccbbadd9-c949-43aa-8d63-af3868304f5d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.001627 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ccbbadd9-c949-43aa-8d63-af3868304f5d" (UID: "ccbbadd9-c949-43aa-8d63-af3868304f5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.027720 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.063998 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.064040 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ccbbadd9-c949-43aa-8d63-af3868304f5d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.165885 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7hhk\" (UniqueName: \"kubernetes.io/projected/dd8830de-44cb-4015-98f2-2d55ffdccf52-kube-api-access-d7hhk\") pod \"dd8830de-44cb-4015-98f2-2d55ffdccf52\" (UID: \"dd8830de-44cb-4015-98f2-2d55ffdccf52\") " Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.169267 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd8830de-44cb-4015-98f2-2d55ffdccf52-kube-api-access-d7hhk" (OuterVolumeSpecName: "kube-api-access-d7hhk") pod "dd8830de-44cb-4015-98f2-2d55ffdccf52" (UID: "dd8830de-44cb-4015-98f2-2d55ffdccf52"). InnerVolumeSpecName "kube-api-access-d7hhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.175286 4979 generic.go:334] "Generic (PLEG): container finished" podID="ccbbadd9-c949-43aa-8d63-af3868304f5d" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" exitCode=0 Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.175365 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ccbbadd9-c949-43aa-8d63-af3868304f5d","Type":"ContainerDied","Data":"d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158"} Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.175393 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ccbbadd9-c949-43aa-8d63-af3868304f5d","Type":"ContainerDied","Data":"78dc60a4fb3d8c231be6536fb69331ebc8d8b2f459882f6ef63cc69939b12033"} Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.175413 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.175474 4979 scope.go:117] "RemoveContainer" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.182742 4979 generic.go:334] "Generic (PLEG): container finished" podID="dd8830de-44cb-4015-98f2-2d55ffdccf52" containerID="54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036" exitCode=2 Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.182787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"dd8830de-44cb-4015-98f2-2d55ffdccf52","Type":"ContainerDied","Data":"54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036"} Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.182814 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.182820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"dd8830de-44cb-4015-98f2-2d55ffdccf52","Type":"ContainerDied","Data":"d0e20d5ba03db06272c4946472538fa78cca5f30fc71f6b0f192583fd5958252"} Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.213838 4979 scope.go:117] "RemoveContainer" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" Dec 04 12:06:44 crc kubenswrapper[4979]: E1204 12:06:44.215549 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158\": container with ID starting with d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158 not found: ID does not exist" containerID="d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.215604 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158"} err="failed to get container status \"d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158\": rpc error: code = NotFound desc = could not find container \"d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158\": container with ID starting with d1f1277e9c9e407925ad222f9909c49640c57fd9c502bd864cddfeb9def81158 not found: ID does not exist" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.215639 4979 scope.go:117] "RemoveContainer" containerID="54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.264372 4979 scope.go:117] "RemoveContainer" containerID="54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.264435 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: E1204 12:06:44.264965 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036\": container with ID starting with 54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036 not found: ID does not exist" containerID="54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.265042 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036"} err="failed to get container status \"54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036\": rpc error: code = NotFound desc = could not find container \"54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036\": container with ID starting with 54d3a5db1aa7c326697d753875d224d058bbe16594b50d86836e93e603ace036 not found: ID does not exist" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.269622 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7hhk\" (UniqueName: \"kubernetes.io/projected/dd8830de-44cb-4015-98f2-2d55ffdccf52-kube-api-access-d7hhk\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.284277 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.306682 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.315879 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: E1204 12:06:44.316348 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccbbadd9-c949-43aa-8d63-af3868304f5d" containerName="nova-scheduler-scheduler" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.316367 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccbbadd9-c949-43aa-8d63-af3868304f5d" containerName="nova-scheduler-scheduler" Dec 04 12:06:44 crc kubenswrapper[4979]: E1204 12:06:44.316390 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd8830de-44cb-4015-98f2-2d55ffdccf52" containerName="kube-state-metrics" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.316397 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd8830de-44cb-4015-98f2-2d55ffdccf52" containerName="kube-state-metrics" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.316575 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccbbadd9-c949-43aa-8d63-af3868304f5d" containerName="nova-scheduler-scheduler" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.316606 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd8830de-44cb-4015-98f2-2d55ffdccf52" containerName="kube-state-metrics" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.317187 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.318939 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.324560 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.335551 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.346052 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.347233 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.350317 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.350496 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.355066 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.472873 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d6dk\" (UniqueName: \"kubernetes.io/projected/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-api-access-6d6dk\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.472952 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.473030 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.473050 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.473173 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf2kn\" (UniqueName: \"kubernetes.io/projected/428b56cd-9aab-4b8b-a54b-f6198243c61a-kube-api-access-vf2kn\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.473214 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-config-data\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.473286 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.568515 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.568564 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.574766 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.574820 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.574916 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf2kn\" (UniqueName: \"kubernetes.io/projected/428b56cd-9aab-4b8b-a54b-f6198243c61a-kube-api-access-vf2kn\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.574948 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-config-data\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.575007 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.575046 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d6dk\" (UniqueName: \"kubernetes.io/projected/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-api-access-6d6dk\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.575155 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.579032 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.579086 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.579209 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-config-data\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.579252 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.580064 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.593011 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf2kn\" (UniqueName: \"kubernetes.io/projected/428b56cd-9aab-4b8b-a54b-f6198243c61a-kube-api-access-vf2kn\") pod \"nova-scheduler-0\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.596652 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d6dk\" (UniqueName: \"kubernetes.io/projected/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-api-access-6d6dk\") pod \"kube-state-metrics-0\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " pod="openstack/kube-state-metrics-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.647670 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:06:44 crc kubenswrapper[4979]: I1204 12:06:44.662443 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.116864 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: W1204 12:06:45.120839 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod428b56cd_9aab_4b8b_a54b_f6198243c61a.slice/crio-8cc55711023546d2237189b16cf0230233b270e0be186238c95c866a93b359e1 WatchSource:0}: Error finding container 8cc55711023546d2237189b16cf0230233b270e0be186238c95c866a93b359e1: Status 404 returned error can't find the container with id 8cc55711023546d2237189b16cf0230233b270e0be186238c95c866a93b359e1 Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.200857 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.209140 4979 generic.go:334] "Generic (PLEG): container finished" podID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerID="93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744" exitCode=0 Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.209236 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7167421-ea96-46d2-882d-a4cabc2e4909","Type":"ContainerDied","Data":"93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744"} Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.209261 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d7167421-ea96-46d2-882d-a4cabc2e4909","Type":"ContainerDied","Data":"1d7cfd21c6180eb5b4b97d9d1112cc76a32d9463e94463e379080f796818511e"} Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.209281 4979 scope.go:117] "RemoveContainer" containerID="93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.209449 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.213050 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"428b56cd-9aab-4b8b-a54b-f6198243c61a","Type":"ContainerStarted","Data":"8cc55711023546d2237189b16cf0230233b270e0be186238c95c866a93b359e1"} Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.215942 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.256470 4979 scope.go:117] "RemoveContainer" containerID="25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.276999 4979 scope.go:117] "RemoveContainer" containerID="93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744" Dec 04 12:06:45 crc kubenswrapper[4979]: E1204 12:06:45.281392 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744\": container with ID starting with 93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744 not found: ID does not exist" containerID="93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.281439 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744"} err="failed to get container status \"93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744\": rpc error: code = NotFound desc = could not find container \"93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744\": container with ID starting with 93374d68f1d9ad41f2f6f1828e2db13541429eb8eb3aa86102b7b16e100f5744 not found: ID does not exist" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.281465 4979 scope.go:117] "RemoveContainer" containerID="25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae" Dec 04 12:06:45 crc kubenswrapper[4979]: E1204 12:06:45.282393 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae\": container with ID starting with 25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae not found: ID does not exist" containerID="25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.282432 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae"} err="failed to get container status \"25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae\": rpc error: code = NotFound desc = could not find container \"25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae\": container with ID starting with 25ba4e41c17fad494b88043e46cc0e983e1894a2dac177e51b513966535e30ae not found: ID does not exist" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.299931 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-config-data\") pod \"d7167421-ea96-46d2-882d-a4cabc2e4909\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.299976 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7167421-ea96-46d2-882d-a4cabc2e4909-logs\") pod \"d7167421-ea96-46d2-882d-a4cabc2e4909\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.300004 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-combined-ca-bundle\") pod \"d7167421-ea96-46d2-882d-a4cabc2e4909\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.300055 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dbtj\" (UniqueName: \"kubernetes.io/projected/d7167421-ea96-46d2-882d-a4cabc2e4909-kube-api-access-2dbtj\") pod \"d7167421-ea96-46d2-882d-a4cabc2e4909\" (UID: \"d7167421-ea96-46d2-882d-a4cabc2e4909\") " Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.301614 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7167421-ea96-46d2-882d-a4cabc2e4909-logs" (OuterVolumeSpecName: "logs") pod "d7167421-ea96-46d2-882d-a4cabc2e4909" (UID: "d7167421-ea96-46d2-882d-a4cabc2e4909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.304054 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7167421-ea96-46d2-882d-a4cabc2e4909-kube-api-access-2dbtj" (OuterVolumeSpecName: "kube-api-access-2dbtj") pod "d7167421-ea96-46d2-882d-a4cabc2e4909" (UID: "d7167421-ea96-46d2-882d-a4cabc2e4909"). InnerVolumeSpecName "kube-api-access-2dbtj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.332359 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-config-data" (OuterVolumeSpecName: "config-data") pod "d7167421-ea96-46d2-882d-a4cabc2e4909" (UID: "d7167421-ea96-46d2-882d-a4cabc2e4909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.333282 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7167421-ea96-46d2-882d-a4cabc2e4909" (UID: "d7167421-ea96-46d2-882d-a4cabc2e4909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.401471 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.401506 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7167421-ea96-46d2-882d-a4cabc2e4909-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.401518 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7167421-ea96-46d2-882d-a4cabc2e4909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.401531 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dbtj\" (UniqueName: \"kubernetes.io/projected/d7167421-ea96-46d2-882d-a4cabc2e4909-kube-api-access-2dbtj\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.568161 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.576460 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.589697 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: E1204 12:06:45.590087 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-api" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.590312 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-api" Dec 04 12:06:45 crc kubenswrapper[4979]: E1204 12:06:45.590337 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-log" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.590347 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-log" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.590559 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-log" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.590593 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" containerName="nova-api-api" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.591859 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.594504 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.604828 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-config-data\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.604954 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35160d93-d21a-40ad-968a-1f06598ad403-logs\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.604994 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx4t4\" (UniqueName: \"kubernetes.io/projected/35160d93-d21a-40ad-968a-1f06598ad403-kube-api-access-qx4t4\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.605014 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.610377 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.713439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-config-data\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.713578 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35160d93-d21a-40ad-968a-1f06598ad403-logs\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.713628 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx4t4\" (UniqueName: \"kubernetes.io/projected/35160d93-d21a-40ad-968a-1f06598ad403-kube-api-access-qx4t4\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.713652 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718077 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718200 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718214 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35160d93-d21a-40ad-968a-1f06598ad403-logs\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718474 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-central-agent" containerID="cri-o://a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697" gracePeriod=30 Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718621 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="proxy-httpd" containerID="cri-o://021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d" gracePeriod=30 Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718686 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="sg-core" containerID="cri-o://da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997" gracePeriod=30 Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.718732 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-notification-agent" containerID="cri-o://5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5" gracePeriod=30 Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.719656 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-config-data\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.737111 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx4t4\" (UniqueName: \"kubernetes.io/projected/35160d93-d21a-40ad-968a-1f06598ad403-kube-api-access-qx4t4\") pod \"nova-api-0\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " pod="openstack/nova-api-0" Dec 04 12:06:45 crc kubenswrapper[4979]: I1204 12:06:45.907334 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.385679 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccbbadd9-c949-43aa-8d63-af3868304f5d" path="/var/lib/kubelet/pods/ccbbadd9-c949-43aa-8d63-af3868304f5d/volumes" Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.388394 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7167421-ea96-46d2-882d-a4cabc2e4909" path="/var/lib/kubelet/pods/d7167421-ea96-46d2-882d-a4cabc2e4909/volumes" Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.388982 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd8830de-44cb-4015-98f2-2d55ffdccf52" path="/var/lib/kubelet/pods/dd8830de-44cb-4015-98f2-2d55ffdccf52/volumes" Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.399867 4979 generic.go:334] "Generic (PLEG): container finished" podID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerID="021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d" exitCode=0 Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.399925 4979 generic.go:334] "Generic (PLEG): container finished" podID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerID="da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997" exitCode=2 Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.399943 4979 generic.go:334] "Generic (PLEG): container finished" podID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerID="a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697" exitCode=0 Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.400109 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerDied","Data":"021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d"} Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.400151 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerDied","Data":"da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997"} Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.400188 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerDied","Data":"a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697"} Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.410868 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"13394b46-eca0-4f42-b3e7-88e20f0bb590","Type":"ContainerStarted","Data":"3300143eb0aae60225a16a5e6c14ef4fb760d477daea629b42dab44e38560ddc"} Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.410909 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"13394b46-eca0-4f42-b3e7-88e20f0bb590","Type":"ContainerStarted","Data":"93d75a52157d4e3ace8193c986b29a1deea9b116c20ff06d193387a34b3650ed"} Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.411910 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.415939 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"428b56cd-9aab-4b8b-a54b-f6198243c61a","Type":"ContainerStarted","Data":"3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5"} Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.481546 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.492486 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.492468201 podStartE2EDuration="2.492468201s" podCreationTimestamp="2025-12-04 12:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:46.476917123 +0000 UTC m=+1430.751212947" watchObservedRunningTime="2025-12-04 12:06:46.492468201 +0000 UTC m=+1430.766763995" Dec 04 12:06:46 crc kubenswrapper[4979]: I1204 12:06:46.501107 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.084338263 podStartE2EDuration="2.501088563s" podCreationTimestamp="2025-12-04 12:06:44 +0000 UTC" firstStartedPulling="2025-12-04 12:06:45.225588914 +0000 UTC m=+1429.499884718" lastFinishedPulling="2025-12-04 12:06:45.642339204 +0000 UTC m=+1429.916635018" observedRunningTime="2025-12-04 12:06:46.493213832 +0000 UTC m=+1430.767509636" watchObservedRunningTime="2025-12-04 12:06:46.501088563 +0000 UTC m=+1430.775384367" Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.426102 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35160d93-d21a-40ad-968a-1f06598ad403","Type":"ContainerStarted","Data":"e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f"} Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.426521 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35160d93-d21a-40ad-968a-1f06598ad403","Type":"ContainerStarted","Data":"0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f"} Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.426533 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35160d93-d21a-40ad-968a-1f06598ad403","Type":"ContainerStarted","Data":"48b93948eb22d40e05306fd27b61cf943863046b97c4677a9b326ded9d572599"} Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.446988 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.446968839 podStartE2EDuration="2.446968839s" podCreationTimestamp="2025-12-04 12:06:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:06:47.442394136 +0000 UTC m=+1431.716689950" watchObservedRunningTime="2025-12-04 12:06:47.446968839 +0000 UTC m=+1431.721264643" Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.949779 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990476 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-config-data\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990549 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-run-httpd\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990577 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-sg-core-conf-yaml\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990598 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-combined-ca-bundle\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990819 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-log-httpd\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990843 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-scripts\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.990912 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fklm9\" (UniqueName: \"kubernetes.io/projected/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-kube-api-access-fklm9\") pod \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\" (UID: \"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b\") " Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.991997 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.992021 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:06:47 crc kubenswrapper[4979]: I1204 12:06:47.997832 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-kube-api-access-fklm9" (OuterVolumeSpecName: "kube-api-access-fklm9") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "kube-api-access-fklm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.000511 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-scripts" (OuterVolumeSpecName: "scripts") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.051708 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.094161 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.094189 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.094198 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fklm9\" (UniqueName: \"kubernetes.io/projected/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-kube-api-access-fklm9\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.094210 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.094219 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.106136 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.116307 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-config-data" (OuterVolumeSpecName: "config-data") pod "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" (UID: "d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.195557 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.195589 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.439286 4979 generic.go:334] "Generic (PLEG): container finished" podID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerID="5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5" exitCode=0 Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.440235 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.440840 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerDied","Data":"5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5"} Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.440873 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b","Type":"ContainerDied","Data":"6e6a7eec92198c13a142cd6e1317018f52a24c211ba7f1be971d545aa8b0923a"} Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.440893 4979 scope.go:117] "RemoveContainer" containerID="021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.465551 4979 scope.go:117] "RemoveContainer" containerID="da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.471969 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.480215 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.487392 4979 scope.go:117] "RemoveContainer" containerID="5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.509510 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.509974 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-notification-agent" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510025 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-notification-agent" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.510053 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="proxy-httpd" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510060 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="proxy-httpd" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.510074 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-central-agent" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510082 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-central-agent" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.510097 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="sg-core" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510105 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="sg-core" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510361 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="proxy-httpd" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510389 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="sg-core" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510407 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-central-agent" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.510420 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" containerName="ceilometer-notification-agent" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.512662 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.516426 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.516447 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.516459 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.517794 4979 scope.go:117] "RemoveContainer" containerID="a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.588228 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602611 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8dfd\" (UniqueName: \"kubernetes.io/projected/4181f9ab-ab56-4131-a458-9d8336bd4b20-kube-api-access-t8dfd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602656 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-config-data\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602694 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602750 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602840 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-scripts\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602865 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602898 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-run-httpd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.602918 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-log-httpd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.603443 4979 scope.go:117] "RemoveContainer" containerID="021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.603935 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d\": container with ID starting with 021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d not found: ID does not exist" containerID="021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.604028 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d"} err="failed to get container status \"021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d\": rpc error: code = NotFound desc = could not find container \"021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d\": container with ID starting with 021801ec0f6141363a70a56c173625470ab3bf1b468bb0fe32c3bffd9df8cd5d not found: ID does not exist" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.604111 4979 scope.go:117] "RemoveContainer" containerID="da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.604522 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997\": container with ID starting with da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997 not found: ID does not exist" containerID="da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.604606 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997"} err="failed to get container status \"da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997\": rpc error: code = NotFound desc = could not find container \"da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997\": container with ID starting with da68a8435c0591463a185a0b9d42779c6aacb5499bb841a90475b8cc4f68e997 not found: ID does not exist" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.604673 4979 scope.go:117] "RemoveContainer" containerID="5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.604938 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5\": container with ID starting with 5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5 not found: ID does not exist" containerID="5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.605027 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5"} err="failed to get container status \"5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5\": rpc error: code = NotFound desc = could not find container \"5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5\": container with ID starting with 5e4ec200fc3233c3ec3318b98e5df7d0d3d3da993d6e33725b9c6c36a5ca46e5 not found: ID does not exist" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.605092 4979 scope.go:117] "RemoveContainer" containerID="a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697" Dec 04 12:06:48 crc kubenswrapper[4979]: E1204 12:06:48.605387 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697\": container with ID starting with a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697 not found: ID does not exist" containerID="a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.605494 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697"} err="failed to get container status \"a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697\": rpc error: code = NotFound desc = could not find container \"a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697\": container with ID starting with a782223db4808cedb61a557f2e8655c160a48c428f176a7509a718b4d9601697 not found: ID does not exist" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704637 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-scripts\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704690 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704732 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-run-httpd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704760 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-log-httpd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704806 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8dfd\" (UniqueName: \"kubernetes.io/projected/4181f9ab-ab56-4131-a458-9d8336bd4b20-kube-api-access-t8dfd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704832 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-config-data\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704876 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.704910 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.705789 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-log-httpd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.706174 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-run-httpd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.708830 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.709204 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.709578 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.712102 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-config-data\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.720125 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-scripts\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.723408 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8dfd\" (UniqueName: \"kubernetes.io/projected/4181f9ab-ab56-4131-a458-9d8336bd4b20-kube-api-access-t8dfd\") pod \"ceilometer-0\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " pod="openstack/ceilometer-0" Dec 04 12:06:48 crc kubenswrapper[4979]: I1204 12:06:48.899269 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.392756 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.393288 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.449353 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerStarted","Data":"ab2a7ebd45e6742a5bfdb5cf95dc2413a102f4faacec785a9e259322a2a9dd74"} Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.552437 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.568668 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.569769 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 12:06:49 crc kubenswrapper[4979]: I1204 12:06:49.648792 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 12:06:50 crc kubenswrapper[4979]: I1204 12:06:50.210802 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b" path="/var/lib/kubelet/pods/d30685c9-3bda-42d2-9a5b-d4ca5a6bcc2b/volumes" Dec 04 12:06:50 crc kubenswrapper[4979]: I1204 12:06:50.461185 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerStarted","Data":"60384221121e39785131cc2e8163801dfee1d9cb4aac2d01abbac8718f644f0f"} Dec 04 12:06:50 crc kubenswrapper[4979]: I1204 12:06:50.582492 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 12:06:50 crc kubenswrapper[4979]: I1204 12:06:50.582837 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 12:06:51 crc kubenswrapper[4979]: I1204 12:06:51.478763 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerStarted","Data":"a3b69e9934ab054f38854c4863aa2601ab7d88a5688a44a733572e1d227494dd"} Dec 04 12:06:52 crc kubenswrapper[4979]: I1204 12:06:52.495740 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerStarted","Data":"5470a0eef3b8ef3cbfe27789e15659e06ce37316ec6761b1c069648a34622f76"} Dec 04 12:06:53 crc kubenswrapper[4979]: I1204 12:06:53.519948 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerStarted","Data":"363e955cac05dca5d386027b3c9305994ce8157bcc68783e5a8ba3bd85137f48"} Dec 04 12:06:53 crc kubenswrapper[4979]: I1204 12:06:53.522273 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:06:53 crc kubenswrapper[4979]: I1204 12:06:53.559014 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.320601102 podStartE2EDuration="5.558991456s" podCreationTimestamp="2025-12-04 12:06:48 +0000 UTC" firstStartedPulling="2025-12-04 12:06:49.393105772 +0000 UTC m=+1433.667401576" lastFinishedPulling="2025-12-04 12:06:52.631496116 +0000 UTC m=+1436.905791930" observedRunningTime="2025-12-04 12:06:53.553800896 +0000 UTC m=+1437.828096720" watchObservedRunningTime="2025-12-04 12:06:53.558991456 +0000 UTC m=+1437.833287270" Dec 04 12:06:54 crc kubenswrapper[4979]: I1204 12:06:54.647941 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 12:06:54 crc kubenswrapper[4979]: I1204 12:06:54.679545 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 12:06:54 crc kubenswrapper[4979]: I1204 12:06:54.686126 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 12:06:55 crc kubenswrapper[4979]: I1204 12:06:55.580614 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 12:06:55 crc kubenswrapper[4979]: I1204 12:06:55.908747 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 12:06:55 crc kubenswrapper[4979]: I1204 12:06:55.908857 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 12:06:56 crc kubenswrapper[4979]: I1204 12:06:56.990503 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 12:06:56 crc kubenswrapper[4979]: I1204 12:06:56.990677 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 12:06:59 crc kubenswrapper[4979]: I1204 12:06:59.574483 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 12:06:59 crc kubenswrapper[4979]: I1204 12:06:59.574983 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 12:06:59 crc kubenswrapper[4979]: I1204 12:06:59.579833 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 12:06:59 crc kubenswrapper[4979]: I1204 12:06:59.580129 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.340895 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.410028 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vhsls"] Dec 04 12:07:03 crc kubenswrapper[4979]: E1204 12:07:03.410427 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.410439 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.410631 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.416964 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.428919 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhsls"] Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.516589 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-config-data\") pod \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.517908 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkcxp\" (UniqueName: \"kubernetes.io/projected/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-kube-api-access-dkcxp\") pod \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.518467 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-combined-ca-bundle\") pod \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\" (UID: \"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af\") " Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.519055 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-catalog-content\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.519138 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcmfr\" (UniqueName: \"kubernetes.io/projected/0ac7d34c-33cf-449f-9089-b314f89407f1-kube-api-access-mcmfr\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.521192 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-utilities\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.537282 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-kube-api-access-dkcxp" (OuterVolumeSpecName: "kube-api-access-dkcxp") pod "bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" (UID: "bb114cda-b4f9-4e6e-b9a1-f34437b5b1af"). InnerVolumeSpecName "kube-api-access-dkcxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.545414 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-config-data" (OuterVolumeSpecName: "config-data") pod "bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" (UID: "bb114cda-b4f9-4e6e-b9a1-f34437b5b1af"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.555423 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" (UID: "bb114cda-b4f9-4e6e-b9a1-f34437b5b1af"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623383 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-utilities\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623798 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-catalog-content\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623841 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcmfr\" (UniqueName: \"kubernetes.io/projected/0ac7d34c-33cf-449f-9089-b314f89407f1-kube-api-access-mcmfr\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623944 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkcxp\" (UniqueName: \"kubernetes.io/projected/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-kube-api-access-dkcxp\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623958 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623968 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.623965 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-utilities\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.624313 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-catalog-content\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.635529 4979 generic.go:334] "Generic (PLEG): container finished" podID="bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" containerID="ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8" exitCode=137 Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.635570 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.635582 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af","Type":"ContainerDied","Data":"ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8"} Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.635615 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bb114cda-b4f9-4e6e-b9a1-f34437b5b1af","Type":"ContainerDied","Data":"453b25852a4752b93720b184ea15bea72967462fa507404d355fc30313bfd88f"} Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.635637 4979 scope.go:117] "RemoveContainer" containerID="ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.640221 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcmfr\" (UniqueName: \"kubernetes.io/projected/0ac7d34c-33cf-449f-9089-b314f89407f1-kube-api-access-mcmfr\") pod \"redhat-operators-vhsls\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.659091 4979 scope.go:117] "RemoveContainer" containerID="ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8" Dec 04 12:07:03 crc kubenswrapper[4979]: E1204 12:07:03.662875 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8\": container with ID starting with ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8 not found: ID does not exist" containerID="ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.662910 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8"} err="failed to get container status \"ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8\": rpc error: code = NotFound desc = could not find container \"ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8\": container with ID starting with ddb5aadd5eae92c3420cde6732ec2bf8c8a918290b81b77a56a38ed0f4da40a8 not found: ID does not exist" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.667856 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.677615 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.690030 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.691964 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.702127 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.725047 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.725718 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.725806 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.758763 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.826375 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mztzj\" (UniqueName: \"kubernetes.io/projected/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-kube-api-access-mztzj\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.826430 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.826535 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.826557 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.826606 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.928697 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mztzj\" (UniqueName: \"kubernetes.io/projected/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-kube-api-access-mztzj\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.928757 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.928858 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.928884 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.928932 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.935139 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.937681 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.938238 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.942664 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:03 crc kubenswrapper[4979]: I1204 12:07:03.948011 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mztzj\" (UniqueName: \"kubernetes.io/projected/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-kube-api-access-mztzj\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.035355 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.220524 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb114cda-b4f9-4e6e-b9a1-f34437b5b1af" path="/var/lib/kubelet/pods/bb114cda-b4f9-4e6e-b9a1-f34437b5b1af/volumes" Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.228848 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vhsls"] Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.493034 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:07:04 crc kubenswrapper[4979]: W1204 12:07:04.543729 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac61c2b8_6dc7_41c5_a2e2_63759d5671b6.slice/crio-10aceaaa15915d853e102c8f274d765fecb48b39e94ae3626fc349a5d04aa448 WatchSource:0}: Error finding container 10aceaaa15915d853e102c8f274d765fecb48b39e94ae3626fc349a5d04aa448: Status 404 returned error can't find the container with id 10aceaaa15915d853e102c8f274d765fecb48b39e94ae3626fc349a5d04aa448 Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.664795 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6","Type":"ContainerStarted","Data":"10aceaaa15915d853e102c8f274d765fecb48b39e94ae3626fc349a5d04aa448"} Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.669989 4979 generic.go:334] "Generic (PLEG): container finished" podID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerID="dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5" exitCode=0 Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.670036 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerDied","Data":"dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5"} Dec 04 12:07:04 crc kubenswrapper[4979]: I1204 12:07:04.670064 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerStarted","Data":"a9b78db4e762b77770b3d9ebd2625694b6d5454a9ffc4a14a07fef79923c4c76"} Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.679834 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6","Type":"ContainerStarted","Data":"676d3e7b21b50404e78713930c4c90e6f133e77d13cd221bda1604c38311a704"} Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.683418 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerStarted","Data":"ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52"} Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.703272 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.703253403 podStartE2EDuration="2.703253403s" podCreationTimestamp="2025-12-04 12:07:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:05.698242858 +0000 UTC m=+1449.972538672" watchObservedRunningTime="2025-12-04 12:07:05.703253403 +0000 UTC m=+1449.977549207" Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.912454 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.914121 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.917390 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 12:07:05 crc kubenswrapper[4979]: I1204 12:07:05.919552 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 12:07:06 crc kubenswrapper[4979]: I1204 12:07:06.698794 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 12:07:06 crc kubenswrapper[4979]: I1204 12:07:06.705050 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 12:07:06 crc kubenswrapper[4979]: I1204 12:07:06.932350 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bnbw6"] Dec 04 12:07:06 crc kubenswrapper[4979]: I1204 12:07:06.934052 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:06 crc kubenswrapper[4979]: I1204 12:07:06.979947 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bnbw6"] Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.000507 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.000555 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.000579 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.000686 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-config\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.000715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.000740 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4rpr\" (UniqueName: \"kubernetes.io/projected/3015ecd7-aacc-41ac-897f-083c4cb5bd62-kube-api-access-m4rpr\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.102129 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.102187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.102213 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.102342 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-config\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.102375 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.102402 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4rpr\" (UniqueName: \"kubernetes.io/projected/3015ecd7-aacc-41ac-897f-083c4cb5bd62-kube-api-access-m4rpr\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.103062 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-swift-storage-0\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.103084 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-sb\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.103103 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-svc\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.103480 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-config\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.103679 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-nb\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.132017 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4rpr\" (UniqueName: \"kubernetes.io/projected/3015ecd7-aacc-41ac-897f-083c4cb5bd62-kube-api-access-m4rpr\") pod \"dnsmasq-dns-89c5cd4d5-bnbw6\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.261648 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.709691 4979 generic.go:334] "Generic (PLEG): container finished" podID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerID="ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52" exitCode=0 Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.709791 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerDied","Data":"ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52"} Dec 04 12:07:07 crc kubenswrapper[4979]: W1204 12:07:07.838487 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3015ecd7_aacc_41ac_897f_083c4cb5bd62.slice/crio-2565c222d0c00fb0c94522511739b7af6698bf9986f7738b23cc3d1066e02bde WatchSource:0}: Error finding container 2565c222d0c00fb0c94522511739b7af6698bf9986f7738b23cc3d1066e02bde: Status 404 returned error can't find the container with id 2565c222d0c00fb0c94522511739b7af6698bf9986f7738b23cc3d1066e02bde Dec 04 12:07:07 crc kubenswrapper[4979]: I1204 12:07:07.838961 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bnbw6"] Dec 04 12:07:08 crc kubenswrapper[4979]: I1204 12:07:08.721100 4979 generic.go:334] "Generic (PLEG): container finished" podID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerID="cd301899a7d93d2b87acbfde23d0f5ef3985c8227fa2a11b27c9b5365050435f" exitCode=0 Dec 04 12:07:08 crc kubenswrapper[4979]: I1204 12:07:08.721496 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" event={"ID":"3015ecd7-aacc-41ac-897f-083c4cb5bd62","Type":"ContainerDied","Data":"cd301899a7d93d2b87acbfde23d0f5ef3985c8227fa2a11b27c9b5365050435f"} Dec 04 12:07:08 crc kubenswrapper[4979]: I1204 12:07:08.721574 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" event={"ID":"3015ecd7-aacc-41ac-897f-083c4cb5bd62","Type":"ContainerStarted","Data":"2565c222d0c00fb0c94522511739b7af6698bf9986f7738b23cc3d1066e02bde"} Dec 04 12:07:08 crc kubenswrapper[4979]: I1204 12:07:08.724916 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerStarted","Data":"98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f"} Dec 04 12:07:08 crc kubenswrapper[4979]: I1204 12:07:08.785996 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vhsls" podStartSLOduration=2.315657388 podStartE2EDuration="5.785973905s" podCreationTimestamp="2025-12-04 12:07:03 +0000 UTC" firstStartedPulling="2025-12-04 12:07:04.672572405 +0000 UTC m=+1448.946868209" lastFinishedPulling="2025-12-04 12:07:08.142888922 +0000 UTC m=+1452.417184726" observedRunningTime="2025-12-04 12:07:08.775090183 +0000 UTC m=+1453.049385987" watchObservedRunningTime="2025-12-04 12:07:08.785973905 +0000 UTC m=+1453.060269709" Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.036388 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.077607 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.077884 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-central-agent" containerID="cri-o://60384221121e39785131cc2e8163801dfee1d9cb4aac2d01abbac8718f644f0f" gracePeriod=30 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.078006 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-notification-agent" containerID="cri-o://a3b69e9934ab054f38854c4863aa2601ab7d88a5688a44a733572e1d227494dd" gracePeriod=30 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.078011 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="sg-core" containerID="cri-o://5470a0eef3b8ef3cbfe27789e15659e06ce37316ec6761b1c069648a34622f76" gracePeriod=30 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.078187 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="proxy-httpd" containerID="cri-o://363e955cac05dca5d386027b3c9305994ce8157bcc68783e5a8ba3bd85137f48" gracePeriod=30 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.178062 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.195:3000/\": read tcp 10.217.0.2:41166->10.217.0.195:3000: read: connection reset by peer" Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.316543 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.735936 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" event={"ID":"3015ecd7-aacc-41ac-897f-083c4cb5bd62","Type":"ContainerStarted","Data":"afd5a98b65f1b147b6118a4045a232bfd6c4bd85bfd36e787d65bf8b8b8468a2"} Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.736090 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.739749 4979 generic.go:334] "Generic (PLEG): container finished" podID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerID="363e955cac05dca5d386027b3c9305994ce8157bcc68783e5a8ba3bd85137f48" exitCode=0 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.739813 4979 generic.go:334] "Generic (PLEG): container finished" podID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerID="5470a0eef3b8ef3cbfe27789e15659e06ce37316ec6761b1c069648a34622f76" exitCode=2 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.739830 4979 generic.go:334] "Generic (PLEG): container finished" podID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerID="60384221121e39785131cc2e8163801dfee1d9cb4aac2d01abbac8718f644f0f" exitCode=0 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.740049 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-log" containerID="cri-o://0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f" gracePeriod=30 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.739781 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerDied","Data":"363e955cac05dca5d386027b3c9305994ce8157bcc68783e5a8ba3bd85137f48"} Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.740181 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerDied","Data":"5470a0eef3b8ef3cbfe27789e15659e06ce37316ec6761b1c069648a34622f76"} Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.740204 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerDied","Data":"60384221121e39785131cc2e8163801dfee1d9cb4aac2d01abbac8718f644f0f"} Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.740280 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-api" containerID="cri-o://e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f" gracePeriod=30 Dec 04 12:07:09 crc kubenswrapper[4979]: I1204 12:07:09.770092 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" podStartSLOduration=3.770069429 podStartE2EDuration="3.770069429s" podCreationTimestamp="2025-12-04 12:07:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:09.768746053 +0000 UTC m=+1454.043041867" watchObservedRunningTime="2025-12-04 12:07:09.770069429 +0000 UTC m=+1454.044365243" Dec 04 12:07:10 crc kubenswrapper[4979]: I1204 12:07:10.768213 4979 generic.go:334] "Generic (PLEG): container finished" podID="35160d93-d21a-40ad-968a-1f06598ad403" containerID="0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f" exitCode=143 Dec 04 12:07:10 crc kubenswrapper[4979]: I1204 12:07:10.768279 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35160d93-d21a-40ad-968a-1f06598ad403","Type":"ContainerDied","Data":"0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f"} Dec 04 12:07:12 crc kubenswrapper[4979]: I1204 12:07:12.791200 4979 generic.go:334] "Generic (PLEG): container finished" podID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerID="a3b69e9934ab054f38854c4863aa2601ab7d88a5688a44a733572e1d227494dd" exitCode=0 Dec 04 12:07:12 crc kubenswrapper[4979]: I1204 12:07:12.791279 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerDied","Data":"a3b69e9934ab054f38854c4863aa2601ab7d88a5688a44a733572e1d227494dd"} Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.460077 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638093 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-log-httpd\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638239 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8dfd\" (UniqueName: \"kubernetes.io/projected/4181f9ab-ab56-4131-a458-9d8336bd4b20-kube-api-access-t8dfd\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638388 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-sg-core-conf-yaml\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638428 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-config-data\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638475 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-combined-ca-bundle\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638520 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-run-httpd\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638542 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-scripts\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638560 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-ceilometer-tls-certs\") pod \"4181f9ab-ab56-4131-a458-9d8336bd4b20\" (UID: \"4181f9ab-ab56-4131-a458-9d8336bd4b20\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.638929 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.639127 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.644995 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4181f9ab-ab56-4131-a458-9d8336bd4b20-kube-api-access-t8dfd" (OuterVolumeSpecName: "kube-api-access-t8dfd") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "kube-api-access-t8dfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.645006 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-scripts" (OuterVolumeSpecName: "scripts") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.649736 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.720161 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.739386 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.741274 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.741315 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.741325 4979 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.741339 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.741349 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4181f9ab-ab56-4131-a458-9d8336bd4b20-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.741359 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8dfd\" (UniqueName: \"kubernetes.io/projected/4181f9ab-ab56-4131-a458-9d8336bd4b20-kube-api-access-t8dfd\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.775568 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.775633 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.787057 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-config-data" (OuterVolumeSpecName: "config-data") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.807507 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4181f9ab-ab56-4131-a458-9d8336bd4b20" (UID: "4181f9ab-ab56-4131-a458-9d8336bd4b20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.811515 4979 generic.go:334] "Generic (PLEG): container finished" podID="35160d93-d21a-40ad-968a-1f06598ad403" containerID="e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f" exitCode=0 Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.811597 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.811617 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35160d93-d21a-40ad-968a-1f06598ad403","Type":"ContainerDied","Data":"e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f"} Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.811659 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"35160d93-d21a-40ad-968a-1f06598ad403","Type":"ContainerDied","Data":"48b93948eb22d40e05306fd27b61cf943863046b97c4677a9b326ded9d572599"} Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.811681 4979 scope.go:117] "RemoveContainer" containerID="e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.817258 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4181f9ab-ab56-4131-a458-9d8336bd4b20","Type":"ContainerDied","Data":"ab2a7ebd45e6742a5bfdb5cf95dc2413a102f4faacec785a9e259322a2a9dd74"} Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.817437 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.843284 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-config-data\") pod \"35160d93-d21a-40ad-968a-1f06598ad403\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.843502 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35160d93-d21a-40ad-968a-1f06598ad403-logs\") pod \"35160d93-d21a-40ad-968a-1f06598ad403\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.843573 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qx4t4\" (UniqueName: \"kubernetes.io/projected/35160d93-d21a-40ad-968a-1f06598ad403-kube-api-access-qx4t4\") pod \"35160d93-d21a-40ad-968a-1f06598ad403\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.843633 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-combined-ca-bundle\") pod \"35160d93-d21a-40ad-968a-1f06598ad403\" (UID: \"35160d93-d21a-40ad-968a-1f06598ad403\") " Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.844159 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.844661 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4181f9ab-ab56-4131-a458-9d8336bd4b20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.847438 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35160d93-d21a-40ad-968a-1f06598ad403-logs" (OuterVolumeSpecName: "logs") pod "35160d93-d21a-40ad-968a-1f06598ad403" (UID: "35160d93-d21a-40ad-968a-1f06598ad403"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.850511 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35160d93-d21a-40ad-968a-1f06598ad403-kube-api-access-qx4t4" (OuterVolumeSpecName: "kube-api-access-qx4t4") pod "35160d93-d21a-40ad-968a-1f06598ad403" (UID: "35160d93-d21a-40ad-968a-1f06598ad403"). InnerVolumeSpecName "kube-api-access-qx4t4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.850766 4979 scope.go:117] "RemoveContainer" containerID="0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.862958 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.879243 4979 scope.go:117] "RemoveContainer" containerID="e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.879733 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f\": container with ID starting with e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f not found: ID does not exist" containerID="e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.879770 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f"} err="failed to get container status \"e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f\": rpc error: code = NotFound desc = could not find container \"e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f\": container with ID starting with e17e5209864a1beb12d5e7bcd6e4cbfef22c46efd406a0232c299ff05369650f not found: ID does not exist" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.879801 4979 scope.go:117] "RemoveContainer" containerID="0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.880063 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f\": container with ID starting with 0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f not found: ID does not exist" containerID="0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.880089 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f"} err="failed to get container status \"0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f\": rpc error: code = NotFound desc = could not find container \"0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f\": container with ID starting with 0f0915fc1c0866d4170ead514046ee2b67c6624f56ede374a2818bec7c1ba41f not found: ID does not exist" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.880107 4979 scope.go:117] "RemoveContainer" containerID="363e955cac05dca5d386027b3c9305994ce8157bcc68783e5a8ba3bd85137f48" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.882370 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.888569 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-config-data" (OuterVolumeSpecName: "config-data") pod "35160d93-d21a-40ad-968a-1f06598ad403" (UID: "35160d93-d21a-40ad-968a-1f06598ad403"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.899807 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.900566 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "35160d93-d21a-40ad-968a-1f06598ad403" (UID: "35160d93-d21a-40ad-968a-1f06598ad403"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.900900 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-central-agent" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.900969 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-central-agent" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.901271 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="proxy-httpd" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.901381 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="proxy-httpd" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.901539 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="sg-core" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.901612 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="sg-core" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.901683 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-log" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.901733 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-log" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.901798 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-notification-agent" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.901857 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-notification-agent" Dec 04 12:07:13 crc kubenswrapper[4979]: E1204 12:07:13.901923 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-api" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.901980 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-api" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.902205 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-central-agent" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.902281 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-log" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.902359 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="35160d93-d21a-40ad-968a-1f06598ad403" containerName="nova-api-api" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.902426 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="sg-core" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.902484 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="proxy-httpd" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.902546 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" containerName="ceilometer-notification-agent" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.901414 4979 scope.go:117] "RemoveContainer" containerID="5470a0eef3b8ef3cbfe27789e15659e06ce37316ec6761b1c069648a34622f76" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.904468 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.908504 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.911004 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.911059 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.923432 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.926080 4979 scope.go:117] "RemoveContainer" containerID="a3b69e9934ab054f38854c4863aa2601ab7d88a5688a44a733572e1d227494dd" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.946125 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.946154 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/35160d93-d21a-40ad-968a-1f06598ad403-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.946164 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qx4t4\" (UniqueName: \"kubernetes.io/projected/35160d93-d21a-40ad-968a-1f06598ad403-kube-api-access-qx4t4\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.946175 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/35160d93-d21a-40ad-968a-1f06598ad403-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:13 crc kubenswrapper[4979]: I1204 12:07:13.949077 4979 scope.go:117] "RemoveContainer" containerID="60384221121e39785131cc2e8163801dfee1d9cb4aac2d01abbac8718f644f0f" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.036141 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.047644 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.047952 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-log-httpd\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.048123 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-config-data\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.048275 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hph5n\" (UniqueName: \"kubernetes.io/projected/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-kube-api-access-hph5n\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.048402 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-run-httpd\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.048521 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.048674 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.048802 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-scripts\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.058567 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.145071 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151048 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151098 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-log-httpd\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151153 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-config-data\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151208 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hph5n\" (UniqueName: \"kubernetes.io/projected/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-kube-api-access-hph5n\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151242 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-run-httpd\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151272 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151312 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151333 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-scripts\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151876 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-run-httpd\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.151943 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-log-httpd\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.154179 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.155027 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.155522 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-config-data\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.156603 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-scripts\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.158100 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.166588 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.168170 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.172763 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.173040 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.173038 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.173218 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hph5n\" (UniqueName: \"kubernetes.io/projected/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-kube-api-access-hph5n\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.174535 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.191282 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.225734 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35160d93-d21a-40ad-968a-1f06598ad403" path="/var/lib/kubelet/pods/35160d93-d21a-40ad-968a-1f06598ad403/volumes" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.226755 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4181f9ab-ab56-4131-a458-9d8336bd4b20" path="/var/lib/kubelet/pods/4181f9ab-ab56-4131-a458-9d8336bd4b20/volumes" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.227887 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.355057 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-public-tls-certs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.355173 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.355201 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr8tv\" (UniqueName: \"kubernetes.io/projected/a82dde59-cd58-470b-add6-78829a8adb28-kube-api-access-qr8tv\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.355270 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82dde59-cd58-470b-add6-78829a8adb28-logs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.355357 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-config-data\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.355467 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.456924 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.457264 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-public-tls-certs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.457347 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.457376 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr8tv\" (UniqueName: \"kubernetes.io/projected/a82dde59-cd58-470b-add6-78829a8adb28-kube-api-access-qr8tv\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.457452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82dde59-cd58-470b-add6-78829a8adb28-logs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.457517 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-config-data\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.459639 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82dde59-cd58-470b-add6-78829a8adb28-logs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.463000 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-public-tls-certs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.463228 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-internal-tls-certs\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.464784 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.477343 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr8tv\" (UniqueName: \"kubernetes.io/projected/a82dde59-cd58-470b-add6-78829a8adb28-kube-api-access-qr8tv\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.484555 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-config-data\") pod \"nova-api-0\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.487500 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.688678 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.828854 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerStarted","Data":"c6962f8310b722640e160e37844cd071c65e0d294c02e5b14021d5f63c753d4b"} Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.844791 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vhsls" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="registry-server" probeResult="failure" output=< Dec 04 12:07:14 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 12:07:14 crc kubenswrapper[4979]: > Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.849636 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:07:14 crc kubenswrapper[4979]: I1204 12:07:14.993691 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:14 crc kubenswrapper[4979]: W1204 12:07:14.994684 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda82dde59_cd58_470b_add6_78829a8adb28.slice/crio-9c44f9c36b575b6ed751d5b86c397477276e642fe69d01ba7ccfd231067a5312 WatchSource:0}: Error finding container 9c44f9c36b575b6ed751d5b86c397477276e642fe69d01ba7ccfd231067a5312: Status 404 returned error can't find the container with id 9c44f9c36b575b6ed751d5b86c397477276e642fe69d01ba7ccfd231067a5312 Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.132860 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-cxc6d"] Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.134091 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.136709 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.136758 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.151523 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-cxc6d"] Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.273914 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcxn4\" (UniqueName: \"kubernetes.io/projected/0175c40e-2589-44d3-adec-95479286dedc-kube-api-access-wcxn4\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.274231 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-config-data\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.274326 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-scripts\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.274423 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.375957 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcxn4\" (UniqueName: \"kubernetes.io/projected/0175c40e-2589-44d3-adec-95479286dedc-kube-api-access-wcxn4\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.376025 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-config-data\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.376075 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-scripts\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.376137 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.385702 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-scripts\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.387249 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.387372 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-config-data\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.408220 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcxn4\" (UniqueName: \"kubernetes.io/projected/0175c40e-2589-44d3-adec-95479286dedc-kube-api-access-wcxn4\") pod \"nova-cell1-cell-mapping-cxc6d\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.456522 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.848211 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a82dde59-cd58-470b-add6-78829a8adb28","Type":"ContainerStarted","Data":"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0"} Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.848501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a82dde59-cd58-470b-add6-78829a8adb28","Type":"ContainerStarted","Data":"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5"} Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.848513 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a82dde59-cd58-470b-add6-78829a8adb28","Type":"ContainerStarted","Data":"9c44f9c36b575b6ed751d5b86c397477276e642fe69d01ba7ccfd231067a5312"} Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.856702 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerStarted","Data":"2d030700ed1d0dceefff7c85f3b95944594dfdf24ef4f4fa9ee05c4e4c775576"} Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.881474 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.881455799 podStartE2EDuration="1.881455799s" podCreationTimestamp="2025-12-04 12:07:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:15.874707608 +0000 UTC m=+1460.149003412" watchObservedRunningTime="2025-12-04 12:07:15.881455799 +0000 UTC m=+1460.155751603" Dec 04 12:07:15 crc kubenswrapper[4979]: I1204 12:07:15.925422 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-cxc6d"] Dec 04 12:07:15 crc kubenswrapper[4979]: W1204 12:07:15.928788 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0175c40e_2589_44d3_adec_95479286dedc.slice/crio-9ce6d3655f6370cbcd69c3766d8e08ba1bc0688d9bdeef97cb5136ed5d81e41e WatchSource:0}: Error finding container 9ce6d3655f6370cbcd69c3766d8e08ba1bc0688d9bdeef97cb5136ed5d81e41e: Status 404 returned error can't find the container with id 9ce6d3655f6370cbcd69c3766d8e08ba1bc0688d9bdeef97cb5136ed5d81e41e Dec 04 12:07:16 crc kubenswrapper[4979]: I1204 12:07:16.886755 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerStarted","Data":"c4914f15e8bbbd431d76efb3ff6d2212902c6e00ddaf95b123156178bfa53c1c"} Dec 04 12:07:16 crc kubenswrapper[4979]: I1204 12:07:16.895031 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cxc6d" event={"ID":"0175c40e-2589-44d3-adec-95479286dedc","Type":"ContainerStarted","Data":"4d858ae77dd1a31f97b848851105d4e11d7f14e77dcd0ed16e08b7acf7736bf5"} Dec 04 12:07:16 crc kubenswrapper[4979]: I1204 12:07:16.895072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cxc6d" event={"ID":"0175c40e-2589-44d3-adec-95479286dedc","Type":"ContainerStarted","Data":"9ce6d3655f6370cbcd69c3766d8e08ba1bc0688d9bdeef97cb5136ed5d81e41e"} Dec 04 12:07:16 crc kubenswrapper[4979]: I1204 12:07:16.921434 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-cxc6d" podStartSLOduration=1.921415957 podStartE2EDuration="1.921415957s" podCreationTimestamp="2025-12-04 12:07:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:16.91742426 +0000 UTC m=+1461.191720074" watchObservedRunningTime="2025-12-04 12:07:16.921415957 +0000 UTC m=+1461.195711761" Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.264061 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.387868 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q5shf"] Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.388170 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="dnsmasq-dns" containerID="cri-o://f17eef864c09a58ad72a34fd9c6c3191ef45028bc4ee2edec5f5297016f9e2a8" gracePeriod=10 Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.902370 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerStarted","Data":"488ba15e74f2c14428d32f7d8b8d5f64b28117aabdfcb4b764a8f0e0a1923d0f"} Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.904830 4979 generic.go:334] "Generic (PLEG): container finished" podID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerID="f17eef864c09a58ad72a34fd9c6c3191ef45028bc4ee2edec5f5297016f9e2a8" exitCode=0 Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.904938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" event={"ID":"b31b83e1-6787-4470-94f7-dcdfd9dad5aa","Type":"ContainerDied","Data":"f17eef864c09a58ad72a34fd9c6c3191ef45028bc4ee2edec5f5297016f9e2a8"} Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.905013 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" event={"ID":"b31b83e1-6787-4470-94f7-dcdfd9dad5aa","Type":"ContainerDied","Data":"1abf45c809845bce062ae2dda266dc68c49e8502eaf5a2256db17f698fbff73a"} Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.905030 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1abf45c809845bce062ae2dda266dc68c49e8502eaf5a2256db17f698fbff73a" Dec 04 12:07:17 crc kubenswrapper[4979]: I1204 12:07:17.934235 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.034571 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-nb\") pod \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.034611 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-sb\") pod \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.034664 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-swift-storage-0\") pod \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.034713 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-config\") pod \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.034764 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-svc\") pod \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.034876 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5c2k\" (UniqueName: \"kubernetes.io/projected/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-kube-api-access-d5c2k\") pod \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\" (UID: \"b31b83e1-6787-4470-94f7-dcdfd9dad5aa\") " Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.042536 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-kube-api-access-d5c2k" (OuterVolumeSpecName: "kube-api-access-d5c2k") pod "b31b83e1-6787-4470-94f7-dcdfd9dad5aa" (UID: "b31b83e1-6787-4470-94f7-dcdfd9dad5aa"). InnerVolumeSpecName "kube-api-access-d5c2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.102039 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-config" (OuterVolumeSpecName: "config") pod "b31b83e1-6787-4470-94f7-dcdfd9dad5aa" (UID: "b31b83e1-6787-4470-94f7-dcdfd9dad5aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.113276 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b31b83e1-6787-4470-94f7-dcdfd9dad5aa" (UID: "b31b83e1-6787-4470-94f7-dcdfd9dad5aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.117112 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b31b83e1-6787-4470-94f7-dcdfd9dad5aa" (UID: "b31b83e1-6787-4470-94f7-dcdfd9dad5aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.119801 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b31b83e1-6787-4470-94f7-dcdfd9dad5aa" (UID: "b31b83e1-6787-4470-94f7-dcdfd9dad5aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.135405 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b31b83e1-6787-4470-94f7-dcdfd9dad5aa" (UID: "b31b83e1-6787-4470-94f7-dcdfd9dad5aa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.137425 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.137476 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5c2k\" (UniqueName: \"kubernetes.io/projected/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-kube-api-access-d5c2k\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.137490 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.137498 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.137506 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.137532 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b31b83e1-6787-4470-94f7-dcdfd9dad5aa-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.920474 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerStarted","Data":"c1c754fe678fbfa03811705e43d6a8012c1cb6310eb4b53a2378db11a7f2248c"} Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.920818 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.920542 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.955632 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.628935802 podStartE2EDuration="5.955615872s" podCreationTimestamp="2025-12-04 12:07:13 +0000 UTC" firstStartedPulling="2025-12-04 12:07:14.694229087 +0000 UTC m=+1458.968524891" lastFinishedPulling="2025-12-04 12:07:18.020909157 +0000 UTC m=+1462.295204961" observedRunningTime="2025-12-04 12:07:18.952685293 +0000 UTC m=+1463.226981147" watchObservedRunningTime="2025-12-04 12:07:18.955615872 +0000 UTC m=+1463.229911676" Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.978308 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q5shf"] Dec 04 12:07:18 crc kubenswrapper[4979]: I1204 12:07:18.987443 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-757b4f8459-q5shf"] Dec 04 12:07:20 crc kubenswrapper[4979]: I1204 12:07:20.211293 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" path="/var/lib/kubelet/pods/b31b83e1-6787-4470-94f7-dcdfd9dad5aa/volumes" Dec 04 12:07:20 crc kubenswrapper[4979]: I1204 12:07:20.944328 4979 generic.go:334] "Generic (PLEG): container finished" podID="0175c40e-2589-44d3-adec-95479286dedc" containerID="4d858ae77dd1a31f97b848851105d4e11d7f14e77dcd0ed16e08b7acf7736bf5" exitCode=0 Dec 04 12:07:20 crc kubenswrapper[4979]: I1204 12:07:20.944377 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cxc6d" event={"ID":"0175c40e-2589-44d3-adec-95479286dedc","Type":"ContainerDied","Data":"4d858ae77dd1a31f97b848851105d4e11d7f14e77dcd0ed16e08b7acf7736bf5"} Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.335697 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.425021 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-scripts\") pod \"0175c40e-2589-44d3-adec-95479286dedc\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.425067 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wcxn4\" (UniqueName: \"kubernetes.io/projected/0175c40e-2589-44d3-adec-95479286dedc-kube-api-access-wcxn4\") pod \"0175c40e-2589-44d3-adec-95479286dedc\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.425193 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-config-data\") pod \"0175c40e-2589-44d3-adec-95479286dedc\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.425377 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-combined-ca-bundle\") pod \"0175c40e-2589-44d3-adec-95479286dedc\" (UID: \"0175c40e-2589-44d3-adec-95479286dedc\") " Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.430874 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-scripts" (OuterVolumeSpecName: "scripts") pod "0175c40e-2589-44d3-adec-95479286dedc" (UID: "0175c40e-2589-44d3-adec-95479286dedc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.432007 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0175c40e-2589-44d3-adec-95479286dedc-kube-api-access-wcxn4" (OuterVolumeSpecName: "kube-api-access-wcxn4") pod "0175c40e-2589-44d3-adec-95479286dedc" (UID: "0175c40e-2589-44d3-adec-95479286dedc"). InnerVolumeSpecName "kube-api-access-wcxn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.453757 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0175c40e-2589-44d3-adec-95479286dedc" (UID: "0175c40e-2589-44d3-adec-95479286dedc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.462603 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-config-data" (OuterVolumeSpecName: "config-data") pod "0175c40e-2589-44d3-adec-95479286dedc" (UID: "0175c40e-2589-44d3-adec-95479286dedc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.527134 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.527171 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.527180 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wcxn4\" (UniqueName: \"kubernetes.io/projected/0175c40e-2589-44d3-adec-95479286dedc-kube-api-access-wcxn4\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.527191 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0175c40e-2589-44d3-adec-95479286dedc-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.856250 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-757b4f8459-q5shf" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.187:5353: i/o timeout" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.975349 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-cxc6d" event={"ID":"0175c40e-2589-44d3-adec-95479286dedc","Type":"ContainerDied","Data":"9ce6d3655f6370cbcd69c3766d8e08ba1bc0688d9bdeef97cb5136ed5d81e41e"} Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.975388 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ce6d3655f6370cbcd69c3766d8e08ba1bc0688d9bdeef97cb5136ed5d81e41e" Dec 04 12:07:22 crc kubenswrapper[4979]: I1204 12:07:22.975431 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-cxc6d" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.167036 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.167445 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-log" containerID="cri-o://e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5" gracePeriod=30 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.167573 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-api" containerID="cri-o://21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0" gracePeriod=30 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.183646 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.183851 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="428b56cd-9aab-4b8b-a54b-f6198243c61a" containerName="nova-scheduler-scheduler" containerID="cri-o://3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" gracePeriod=30 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.250428 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.250703 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-log" containerID="cri-o://ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba" gracePeriod=30 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.250786 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-metadata" containerID="cri-o://24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be" gracePeriod=30 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.834212 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.837230 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.886719 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.952240 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr8tv\" (UniqueName: \"kubernetes.io/projected/a82dde59-cd58-470b-add6-78829a8adb28-kube-api-access-qr8tv\") pod \"a82dde59-cd58-470b-add6-78829a8adb28\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.952851 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-public-tls-certs\") pod \"a82dde59-cd58-470b-add6-78829a8adb28\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.953024 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-combined-ca-bundle\") pod \"a82dde59-cd58-470b-add6-78829a8adb28\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.953166 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82dde59-cd58-470b-add6-78829a8adb28-logs\") pod \"a82dde59-cd58-470b-add6-78829a8adb28\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.953493 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-config-data\") pod \"a82dde59-cd58-470b-add6-78829a8adb28\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.953673 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-internal-tls-certs\") pod \"a82dde59-cd58-470b-add6-78829a8adb28\" (UID: \"a82dde59-cd58-470b-add6-78829a8adb28\") " Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.953536 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a82dde59-cd58-470b-add6-78829a8adb28-logs" (OuterVolumeSpecName: "logs") pod "a82dde59-cd58-470b-add6-78829a8adb28" (UID: "a82dde59-cd58-470b-add6-78829a8adb28"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.955006 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a82dde59-cd58-470b-add6-78829a8adb28-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.960612 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a82dde59-cd58-470b-add6-78829a8adb28-kube-api-access-qr8tv" (OuterVolumeSpecName: "kube-api-access-qr8tv") pod "a82dde59-cd58-470b-add6-78829a8adb28" (UID: "a82dde59-cd58-470b-add6-78829a8adb28"). InnerVolumeSpecName "kube-api-access-qr8tv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.984267 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a82dde59-cd58-470b-add6-78829a8adb28" (UID: "a82dde59-cd58-470b-add6-78829a8adb28"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.990259 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-config-data" (OuterVolumeSpecName: "config-data") pod "a82dde59-cd58-470b-add6-78829a8adb28" (UID: "a82dde59-cd58-470b-add6-78829a8adb28"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.993501 4979 generic.go:334] "Generic (PLEG): container finished" podID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerID="ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba" exitCode=143 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.993570 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab","Type":"ContainerDied","Data":"ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba"} Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.995684 4979 generic.go:334] "Generic (PLEG): container finished" podID="a82dde59-cd58-470b-add6-78829a8adb28" containerID="21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0" exitCode=0 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.995703 4979 generic.go:334] "Generic (PLEG): container finished" podID="a82dde59-cd58-470b-add6-78829a8adb28" containerID="e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5" exitCode=143 Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.995835 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.996413 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a82dde59-cd58-470b-add6-78829a8adb28","Type":"ContainerDied","Data":"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0"} Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.996457 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a82dde59-cd58-470b-add6-78829a8adb28","Type":"ContainerDied","Data":"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5"} Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.996472 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"a82dde59-cd58-470b-add6-78829a8adb28","Type":"ContainerDied","Data":"9c44f9c36b575b6ed751d5b86c397477276e642fe69d01ba7ccfd231067a5312"} Dec 04 12:07:23 crc kubenswrapper[4979]: I1204 12:07:23.996491 4979 scope.go:117] "RemoveContainer" containerID="21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.003895 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a82dde59-cd58-470b-add6-78829a8adb28" (UID: "a82dde59-cd58-470b-add6-78829a8adb28"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.019180 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "a82dde59-cd58-470b-add6-78829a8adb28" (UID: "a82dde59-cd58-470b-add6-78829a8adb28"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.042053 4979 scope.go:117] "RemoveContainer" containerID="e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.057385 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.057550 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.057644 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.057716 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a82dde59-cd58-470b-add6-78829a8adb28-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.057783 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr8tv\" (UniqueName: \"kubernetes.io/projected/a82dde59-cd58-470b-add6-78829a8adb28-kube-api-access-qr8tv\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.060197 4979 scope.go:117] "RemoveContainer" containerID="21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.060804 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0\": container with ID starting with 21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0 not found: ID does not exist" containerID="21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.060850 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0"} err="failed to get container status \"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0\": rpc error: code = NotFound desc = could not find container \"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0\": container with ID starting with 21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0 not found: ID does not exist" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.060878 4979 scope.go:117] "RemoveContainer" containerID="e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.061211 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5\": container with ID starting with e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5 not found: ID does not exist" containerID="e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.061320 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5"} err="failed to get container status \"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5\": rpc error: code = NotFound desc = could not find container \"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5\": container with ID starting with e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5 not found: ID does not exist" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.061414 4979 scope.go:117] "RemoveContainer" containerID="21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.061847 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0"} err="failed to get container status \"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0\": rpc error: code = NotFound desc = could not find container \"21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0\": container with ID starting with 21bdba7c5f5df15ed646db6b011f259760cfd25818c6a15d686aefadcdff68e0 not found: ID does not exist" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.061934 4979 scope.go:117] "RemoveContainer" containerID="e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.062249 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5"} err="failed to get container status \"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5\": rpc error: code = NotFound desc = could not find container \"e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5\": container with ID starting with e96ae361b832181590965de7242feceac639c9fb61c21d3014942af1b2901bc5 not found: ID does not exist" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.096952 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhsls"] Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.356042 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.366615 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395166 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.395596 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-api" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395613 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-api" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.395621 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0175c40e-2589-44d3-adec-95479286dedc" containerName="nova-manage" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395626 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0175c40e-2589-44d3-adec-95479286dedc" containerName="nova-manage" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.395641 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="dnsmasq-dns" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395647 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="dnsmasq-dns" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.395674 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="init" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395679 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="init" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.395694 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-log" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395700 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-log" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395864 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b31b83e1-6787-4470-94f7-dcdfd9dad5aa" containerName="dnsmasq-dns" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395882 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-api" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395892 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0175c40e-2589-44d3-adec-95479286dedc" containerName="nova-manage" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.395905 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a82dde59-cd58-470b-add6-78829a8adb28" containerName="nova-api-log" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.397086 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.400932 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.401013 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.401233 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.404704 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.567283 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c472e576-897a-4ea6-912a-2d87b2f67077-logs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.567401 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.567516 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-public-tls-certs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.567581 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-config-data\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.567665 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65zkd\" (UniqueName: \"kubernetes.io/projected/c472e576-897a-4ea6-912a-2d87b2f67077-kube-api-access-65zkd\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.567746 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.650594 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.652165 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.653593 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:07:24 crc kubenswrapper[4979]: E1204 12:07:24.653631 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="428b56cd-9aab-4b8b-a54b-f6198243c61a" containerName="nova-scheduler-scheduler" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.669526 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-config-data\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.669601 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65zkd\" (UniqueName: \"kubernetes.io/projected/c472e576-897a-4ea6-912a-2d87b2f67077-kube-api-access-65zkd\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.669654 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.669688 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c472e576-897a-4ea6-912a-2d87b2f67077-logs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.669712 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.669751 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-public-tls-certs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.670476 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c472e576-897a-4ea6-912a-2d87b2f67077-logs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.673754 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-internal-tls-certs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.674617 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-config-data\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.674707 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-public-tls-certs\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.676015 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.685492 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65zkd\" (UniqueName: \"kubernetes.io/projected/c472e576-897a-4ea6-912a-2d87b2f67077-kube-api-access-65zkd\") pod \"nova-api-0\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " pod="openstack/nova-api-0" Dec 04 12:07:24 crc kubenswrapper[4979]: I1204 12:07:24.715726 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.005040 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vhsls" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="registry-server" containerID="cri-o://98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f" gracePeriod=2 Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.195683 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:07:25 crc kubenswrapper[4979]: W1204 12:07:25.206671 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc472e576_897a_4ea6_912a_2d87b2f67077.slice/crio-5b251dc57aafe3af2839d9c2c0ebaecf76adeef682678c03c10153636b5311df WatchSource:0}: Error finding container 5b251dc57aafe3af2839d9c2c0ebaecf76adeef682678c03c10153636b5311df: Status 404 returned error can't find the container with id 5b251dc57aafe3af2839d9c2c0ebaecf76adeef682678c03c10153636b5311df Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.573882 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.689496 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-utilities\") pod \"0ac7d34c-33cf-449f-9089-b314f89407f1\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.689539 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcmfr\" (UniqueName: \"kubernetes.io/projected/0ac7d34c-33cf-449f-9089-b314f89407f1-kube-api-access-mcmfr\") pod \"0ac7d34c-33cf-449f-9089-b314f89407f1\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.689651 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-catalog-content\") pod \"0ac7d34c-33cf-449f-9089-b314f89407f1\" (UID: \"0ac7d34c-33cf-449f-9089-b314f89407f1\") " Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.690532 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-utilities" (OuterVolumeSpecName: "utilities") pod "0ac7d34c-33cf-449f-9089-b314f89407f1" (UID: "0ac7d34c-33cf-449f-9089-b314f89407f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.694102 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ac7d34c-33cf-449f-9089-b314f89407f1-kube-api-access-mcmfr" (OuterVolumeSpecName: "kube-api-access-mcmfr") pod "0ac7d34c-33cf-449f-9089-b314f89407f1" (UID: "0ac7d34c-33cf-449f-9089-b314f89407f1"). InnerVolumeSpecName "kube-api-access-mcmfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.791586 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.791619 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcmfr\" (UniqueName: \"kubernetes.io/projected/0ac7d34c-33cf-449f-9089-b314f89407f1-kube-api-access-mcmfr\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.829725 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ac7d34c-33cf-449f-9089-b314f89407f1" (UID: "0ac7d34c-33cf-449f-9089-b314f89407f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:25 crc kubenswrapper[4979]: I1204 12:07:25.893502 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ac7d34c-33cf-449f-9089-b314f89407f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.018255 4979 generic.go:334] "Generic (PLEG): container finished" podID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerID="98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f" exitCode=0 Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.018346 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerDied","Data":"98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f"} Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.018348 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vhsls" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.018390 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vhsls" event={"ID":"0ac7d34c-33cf-449f-9089-b314f89407f1","Type":"ContainerDied","Data":"a9b78db4e762b77770b3d9ebd2625694b6d5454a9ffc4a14a07fef79923c4c76"} Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.018410 4979 scope.go:117] "RemoveContainer" containerID="98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.020485 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c472e576-897a-4ea6-912a-2d87b2f67077","Type":"ContainerStarted","Data":"0d566502255055e35e3f3b677f5eefb5b5ec440a9ad5528a85688965a96d78c3"} Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.020545 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c472e576-897a-4ea6-912a-2d87b2f67077","Type":"ContainerStarted","Data":"ea884876a5d807dfae591e01b960188f749527339aa8424bc250430bc801a04f"} Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.020559 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c472e576-897a-4ea6-912a-2d87b2f67077","Type":"ContainerStarted","Data":"5b251dc57aafe3af2839d9c2c0ebaecf76adeef682678c03c10153636b5311df"} Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.055658 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.055640809 podStartE2EDuration="2.055640809s" podCreationTimestamp="2025-12-04 12:07:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:26.054240061 +0000 UTC m=+1470.328535885" watchObservedRunningTime="2025-12-04 12:07:26.055640809 +0000 UTC m=+1470.329936613" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.073170 4979 scope.go:117] "RemoveContainer" containerID="ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.078140 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vhsls"] Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.088034 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vhsls"] Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.093642 4979 scope.go:117] "RemoveContainer" containerID="dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.113330 4979 scope.go:117] "RemoveContainer" containerID="98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f" Dec 04 12:07:26 crc kubenswrapper[4979]: E1204 12:07:26.113811 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f\": container with ID starting with 98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f not found: ID does not exist" containerID="98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.113879 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f"} err="failed to get container status \"98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f\": rpc error: code = NotFound desc = could not find container \"98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f\": container with ID starting with 98b75c1929d851cde55c948577ba8f3cce83ea011186dde0a7c8a6ce59c3e42f not found: ID does not exist" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.113937 4979 scope.go:117] "RemoveContainer" containerID="ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52" Dec 04 12:07:26 crc kubenswrapper[4979]: E1204 12:07:26.114373 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52\": container with ID starting with ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52 not found: ID does not exist" containerID="ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.114466 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52"} err="failed to get container status \"ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52\": rpc error: code = NotFound desc = could not find container \"ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52\": container with ID starting with ff3b8be9dfff5524bb4a0018dce5b2b50e61801aa50ba79fb63fbca1479f1a52 not found: ID does not exist" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.114557 4979 scope.go:117] "RemoveContainer" containerID="dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5" Dec 04 12:07:26 crc kubenswrapper[4979]: E1204 12:07:26.114963 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5\": container with ID starting with dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5 not found: ID does not exist" containerID="dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.115020 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5"} err="failed to get container status \"dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5\": rpc error: code = NotFound desc = could not find container \"dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5\": container with ID starting with dc6cf280c50e72a11af444b3e8cc3989297e82673c8b2329f62f0da1a58099c5 not found: ID does not exist" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.211988 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" path="/var/lib/kubelet/pods/0ac7d34c-33cf-449f-9089-b314f89407f1/volumes" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.212698 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a82dde59-cd58-470b-add6-78829a8adb28" path="/var/lib/kubelet/pods/a82dde59-cd58-470b-add6-78829a8adb28/volumes" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.373771 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:33014->10.217.0.191:8775: read: connection reset by peer" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.373783 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.191:8775/\": read tcp 10.217.0.2:33008->10.217.0.191:8775: read: connection reset by peer" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.812713 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.910053 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-config-data\") pod \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.910379 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zwdr\" (UniqueName: \"kubernetes.io/projected/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-kube-api-access-7zwdr\") pod \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.910502 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-combined-ca-bundle\") pod \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.910646 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-nova-metadata-tls-certs\") pod \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.910815 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-logs\") pod \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\" (UID: \"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab\") " Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.912424 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-logs" (OuterVolumeSpecName: "logs") pod "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" (UID: "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.929017 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-kube-api-access-7zwdr" (OuterVolumeSpecName: "kube-api-access-7zwdr") pod "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" (UID: "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab"). InnerVolumeSpecName "kube-api-access-7zwdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.942016 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-config-data" (OuterVolumeSpecName: "config-data") pod "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" (UID: "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.942438 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" (UID: "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:26 crc kubenswrapper[4979]: I1204 12:07:26.989163 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" (UID: "0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.013505 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.013549 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zwdr\" (UniqueName: \"kubernetes.io/projected/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-kube-api-access-7zwdr\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.013833 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.013852 4979 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.013864 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.035620 4979 generic.go:334] "Generic (PLEG): container finished" podID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerID="24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be" exitCode=0 Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.035673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab","Type":"ContainerDied","Data":"24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be"} Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.035736 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab","Type":"ContainerDied","Data":"da3e0d2b0e1bdedade32c382dfefbbd15d605da522f9f4b7c2c020f985400ecd"} Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.035755 4979 scope.go:117] "RemoveContainer" containerID="24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.035691 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.061498 4979 scope.go:117] "RemoveContainer" containerID="ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.066319 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.089615 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099049 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.099580 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-metadata" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099595 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-metadata" Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.099609 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="extract-content" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099615 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="extract-content" Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.099626 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-log" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099632 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-log" Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.099645 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="extract-utilities" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099651 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="extract-utilities" Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.099665 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="registry-server" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099671 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="registry-server" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099924 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-log" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.099961 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" containerName="nova-metadata-metadata" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.101430 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ac7d34c-33cf-449f-9089-b314f89407f1" containerName="registry-server" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.102759 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.158123 4979 scope.go:117] "RemoveContainer" containerID="24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.158895 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.159167 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be\": container with ID starting with 24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be not found: ID does not exist" containerID="24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.159198 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be"} err="failed to get container status \"24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be\": rpc error: code = NotFound desc = could not find container \"24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be\": container with ID starting with 24bbac0cef6ed3959fe11d411e1567e74d6a68c614cb571c6ca5656f3a91f2be not found: ID does not exist" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.159225 4979 scope.go:117] "RemoveContainer" containerID="ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.159352 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 04 12:07:27 crc kubenswrapper[4979]: E1204 12:07:27.159972 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba\": container with ID starting with ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba not found: ID does not exist" containerID="ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.159994 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba"} err="failed to get container status \"ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba\": rpc error: code = NotFound desc = could not find container \"ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba\": container with ID starting with ed3d3391bbf4d7506a4be054ba025dc615fcde398f25b3939be32dcc997249ba not found: ID does not exist" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.179936 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.264028 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.264083 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.264109 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-config-data\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.264126 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-659ct\" (UniqueName: \"kubernetes.io/projected/3238d4e5-e968-4d66-87d6-7397f48e9714-kube-api-access-659ct\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.264266 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3238d4e5-e968-4d66-87d6-7397f48e9714-logs\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.366374 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.366474 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.366505 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-config-data\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.366529 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-659ct\" (UniqueName: \"kubernetes.io/projected/3238d4e5-e968-4d66-87d6-7397f48e9714-kube-api-access-659ct\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.366758 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3238d4e5-e968-4d66-87d6-7397f48e9714-logs\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.367908 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3238d4e5-e968-4d66-87d6-7397f48e9714-logs\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.370642 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.371472 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.372413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-config-data\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.384836 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-659ct\" (UniqueName: \"kubernetes.io/projected/3238d4e5-e968-4d66-87d6-7397f48e9714-kube-api-access-659ct\") pod \"nova-metadata-0\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " pod="openstack/nova-metadata-0" Dec 04 12:07:27 crc kubenswrapper[4979]: I1204 12:07:27.474380 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.040643 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.041058 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.211739 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab" path="/var/lib/kubelet/pods/0f616ba1-7846-4dc4-93e9-6aae7ae9b4ab/volumes" Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.393141 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.881916 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.997185 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-combined-ca-bundle\") pod \"428b56cd-9aab-4b8b-a54b-f6198243c61a\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.997290 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vf2kn\" (UniqueName: \"kubernetes.io/projected/428b56cd-9aab-4b8b-a54b-f6198243c61a-kube-api-access-vf2kn\") pod \"428b56cd-9aab-4b8b-a54b-f6198243c61a\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " Dec 04 12:07:28 crc kubenswrapper[4979]: I1204 12:07:28.997443 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-config-data\") pod \"428b56cd-9aab-4b8b-a54b-f6198243c61a\" (UID: \"428b56cd-9aab-4b8b-a54b-f6198243c61a\") " Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.002240 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/428b56cd-9aab-4b8b-a54b-f6198243c61a-kube-api-access-vf2kn" (OuterVolumeSpecName: "kube-api-access-vf2kn") pod "428b56cd-9aab-4b8b-a54b-f6198243c61a" (UID: "428b56cd-9aab-4b8b-a54b-f6198243c61a"). InnerVolumeSpecName "kube-api-access-vf2kn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.024356 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-config-data" (OuterVolumeSpecName: "config-data") pod "428b56cd-9aab-4b8b-a54b-f6198243c61a" (UID: "428b56cd-9aab-4b8b-a54b-f6198243c61a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.024889 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "428b56cd-9aab-4b8b-a54b-f6198243c61a" (UID: "428b56cd-9aab-4b8b-a54b-f6198243c61a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.060706 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3238d4e5-e968-4d66-87d6-7397f48e9714","Type":"ContainerStarted","Data":"640bdb1e4d1f6981f6d0fba2f1150c83d60fe36e91add11c47a642a8218b2507"} Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.060748 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3238d4e5-e968-4d66-87d6-7397f48e9714","Type":"ContainerStarted","Data":"601adae13b36c5015c0e550950004699f70cac5db6ce0564490dc4284b7bfc50"} Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.060801 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3238d4e5-e968-4d66-87d6-7397f48e9714","Type":"ContainerStarted","Data":"011026b7a80047db82851700f8e2c0e45e41fd86dc23c3f536e6c524c19a4670"} Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.062202 4979 generic.go:334] "Generic (PLEG): container finished" podID="428b56cd-9aab-4b8b-a54b-f6198243c61a" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" exitCode=0 Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.062238 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"428b56cd-9aab-4b8b-a54b-f6198243c61a","Type":"ContainerDied","Data":"3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5"} Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.062260 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"428b56cd-9aab-4b8b-a54b-f6198243c61a","Type":"ContainerDied","Data":"8cc55711023546d2237189b16cf0230233b270e0be186238c95c866a93b359e1"} Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.062276 4979 scope.go:117] "RemoveContainer" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.062224 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.092978 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.092955999 podStartE2EDuration="2.092955999s" podCreationTimestamp="2025-12-04 12:07:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:29.084705576 +0000 UTC m=+1473.359001390" watchObservedRunningTime="2025-12-04 12:07:29.092955999 +0000 UTC m=+1473.367251793" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.093256 4979 scope.go:117] "RemoveContainer" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" Dec 04 12:07:29 crc kubenswrapper[4979]: E1204 12:07:29.093702 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5\": container with ID starting with 3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5 not found: ID does not exist" containerID="3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.093736 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5"} err="failed to get container status \"3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5\": rpc error: code = NotFound desc = could not find container \"3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5\": container with ID starting with 3ffb26b2ae06a5cf22f8d78eaa713896da2e95a84ee0981b820203146de6edc5 not found: ID does not exist" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.101342 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.101373 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/428b56cd-9aab-4b8b-a54b-f6198243c61a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.101385 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vf2kn\" (UniqueName: \"kubernetes.io/projected/428b56cd-9aab-4b8b-a54b-f6198243c61a-kube-api-access-vf2kn\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.108125 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.123583 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.137100 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:07:29 crc kubenswrapper[4979]: E1204 12:07:29.137662 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="428b56cd-9aab-4b8b-a54b-f6198243c61a" containerName="nova-scheduler-scheduler" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.138060 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="428b56cd-9aab-4b8b-a54b-f6198243c61a" containerName="nova-scheduler-scheduler" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.138317 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="428b56cd-9aab-4b8b-a54b-f6198243c61a" containerName="nova-scheduler-scheduler" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.139079 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.142040 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.148385 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:07:29 crc kubenswrapper[4979]: E1204 12:07:29.153697 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod428b56cd_9aab_4b8b_a54b_f6198243c61a.slice/crio-8cc55711023546d2237189b16cf0230233b270e0be186238c95c866a93b359e1\": RecentStats: unable to find data in memory cache]" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.304766 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.305028 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.305348 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7b7x\" (UniqueName: \"kubernetes.io/projected/77ca261e-57dc-4b57-a80b-e7def9b113d5-kube-api-access-c7b7x\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.406762 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7b7x\" (UniqueName: \"kubernetes.io/projected/77ca261e-57dc-4b57-a80b-e7def9b113d5-kube-api-access-c7b7x\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.406929 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.406985 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.411797 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.412160 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.424693 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7b7x\" (UniqueName: \"kubernetes.io/projected/77ca261e-57dc-4b57-a80b-e7def9b113d5-kube-api-access-c7b7x\") pod \"nova-scheduler-0\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.485902 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:07:29 crc kubenswrapper[4979]: I1204 12:07:29.956589 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:07:29 crc kubenswrapper[4979]: W1204 12:07:29.966915 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod77ca261e_57dc_4b57_a80b_e7def9b113d5.slice/crio-350234e9448f97a3bc3e3fa4d40d902c428f14deb9688f82dadaa360f8fb5930 WatchSource:0}: Error finding container 350234e9448f97a3bc3e3fa4d40d902c428f14deb9688f82dadaa360f8fb5930: Status 404 returned error can't find the container with id 350234e9448f97a3bc3e3fa4d40d902c428f14deb9688f82dadaa360f8fb5930 Dec 04 12:07:30 crc kubenswrapper[4979]: I1204 12:07:30.077613 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77ca261e-57dc-4b57-a80b-e7def9b113d5","Type":"ContainerStarted","Data":"350234e9448f97a3bc3e3fa4d40d902c428f14deb9688f82dadaa360f8fb5930"} Dec 04 12:07:30 crc kubenswrapper[4979]: I1204 12:07:30.212216 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="428b56cd-9aab-4b8b-a54b-f6198243c61a" path="/var/lib/kubelet/pods/428b56cd-9aab-4b8b-a54b-f6198243c61a/volumes" Dec 04 12:07:31 crc kubenswrapper[4979]: I1204 12:07:31.086999 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77ca261e-57dc-4b57-a80b-e7def9b113d5","Type":"ContainerStarted","Data":"f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27"} Dec 04 12:07:31 crc kubenswrapper[4979]: I1204 12:07:31.106465 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.106445986 podStartE2EDuration="2.106445986s" podCreationTimestamp="2025-12-04 12:07:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:07:31.104941985 +0000 UTC m=+1475.379237839" watchObservedRunningTime="2025-12-04 12:07:31.106445986 +0000 UTC m=+1475.380741800" Dec 04 12:07:32 crc kubenswrapper[4979]: I1204 12:07:32.474810 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 12:07:32 crc kubenswrapper[4979]: I1204 12:07:32.474877 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 12:07:34 crc kubenswrapper[4979]: I1204 12:07:34.486716 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 12:07:34 crc kubenswrapper[4979]: I1204 12:07:34.716349 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 12:07:34 crc kubenswrapper[4979]: I1204 12:07:34.716417 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 12:07:35 crc kubenswrapper[4979]: I1204 12:07:35.737587 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 12:07:35 crc kubenswrapper[4979]: I1204 12:07:35.737684 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 12:07:37 crc kubenswrapper[4979]: I1204 12:07:37.475153 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 12:07:37 crc kubenswrapper[4979]: I1204 12:07:37.475211 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 12:07:38 crc kubenswrapper[4979]: I1204 12:07:38.491652 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 12:07:38 crc kubenswrapper[4979]: I1204 12:07:38.491652 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 12:07:39 crc kubenswrapper[4979]: I1204 12:07:39.487044 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 12:07:39 crc kubenswrapper[4979]: I1204 12:07:39.519402 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.039481 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nmrqp"] Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.042596 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.058686 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmrqp"] Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.124537 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-catalog-content\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.124940 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8xmr\" (UniqueName: \"kubernetes.io/projected/b1a72d2d-9728-492f-9e18-0cb188740701-kube-api-access-f8xmr\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.125063 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-utilities\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.227187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-utilities\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.227256 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-catalog-content\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.227325 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8xmr\" (UniqueName: \"kubernetes.io/projected/b1a72d2d-9728-492f-9e18-0cb188740701-kube-api-access-f8xmr\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.227820 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-utilities\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.227854 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-catalog-content\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.248307 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8xmr\" (UniqueName: \"kubernetes.io/projected/b1a72d2d-9728-492f-9e18-0cb188740701-kube-api-access-f8xmr\") pod \"redhat-marketplace-nmrqp\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.249829 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.367154 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:40 crc kubenswrapper[4979]: I1204 12:07:40.874359 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmrqp"] Dec 04 12:07:41 crc kubenswrapper[4979]: I1204 12:07:41.224147 4979 generic.go:334] "Generic (PLEG): container finished" podID="b1a72d2d-9728-492f-9e18-0cb188740701" containerID="166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f" exitCode=0 Dec 04 12:07:41 crc kubenswrapper[4979]: I1204 12:07:41.224213 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmrqp" event={"ID":"b1a72d2d-9728-492f-9e18-0cb188740701","Type":"ContainerDied","Data":"166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f"} Dec 04 12:07:41 crc kubenswrapper[4979]: I1204 12:07:41.224959 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmrqp" event={"ID":"b1a72d2d-9728-492f-9e18-0cb188740701","Type":"ContainerStarted","Data":"6db0ed631a70e7656f2a6aed945b1cffe5305574a0bbd9f7cb8080934c48db6a"} Dec 04 12:07:43 crc kubenswrapper[4979]: I1204 12:07:43.244513 4979 generic.go:334] "Generic (PLEG): container finished" podID="b1a72d2d-9728-492f-9e18-0cb188740701" containerID="240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098" exitCode=0 Dec 04 12:07:43 crc kubenswrapper[4979]: I1204 12:07:43.245905 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmrqp" event={"ID":"b1a72d2d-9728-492f-9e18-0cb188740701","Type":"ContainerDied","Data":"240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098"} Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.236020 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.256063 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmrqp" event={"ID":"b1a72d2d-9728-492f-9e18-0cb188740701","Type":"ContainerStarted","Data":"ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a"} Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.305210 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nmrqp" podStartSLOduration=1.7503168489999998 podStartE2EDuration="4.305189692s" podCreationTimestamp="2025-12-04 12:07:40 +0000 UTC" firstStartedPulling="2025-12-04 12:07:41.226048084 +0000 UTC m=+1485.500343898" lastFinishedPulling="2025-12-04 12:07:43.780920897 +0000 UTC m=+1488.055216741" observedRunningTime="2025-12-04 12:07:44.296633331 +0000 UTC m=+1488.570929145" watchObservedRunningTime="2025-12-04 12:07:44.305189692 +0000 UTC m=+1488.579485496" Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.729026 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.730356 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.730446 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 12:07:44 crc kubenswrapper[4979]: I1204 12:07:44.738868 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 12:07:45 crc kubenswrapper[4979]: I1204 12:07:45.268874 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 12:07:45 crc kubenswrapper[4979]: I1204 12:07:45.413805 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 12:07:47 crc kubenswrapper[4979]: I1204 12:07:47.479293 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 12:07:47 crc kubenswrapper[4979]: I1204 12:07:47.479879 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 12:07:47 crc kubenswrapper[4979]: I1204 12:07:47.484409 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 12:07:47 crc kubenswrapper[4979]: I1204 12:07:47.488057 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 12:07:50 crc kubenswrapper[4979]: I1204 12:07:50.368260 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:50 crc kubenswrapper[4979]: I1204 12:07:50.368888 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:50 crc kubenswrapper[4979]: I1204 12:07:50.429223 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:51 crc kubenswrapper[4979]: I1204 12:07:51.376466 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:51 crc kubenswrapper[4979]: I1204 12:07:51.425251 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmrqp"] Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.332561 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nmrqp" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="registry-server" containerID="cri-o://ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a" gracePeriod=2 Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.737489 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.807070 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-catalog-content\") pod \"b1a72d2d-9728-492f-9e18-0cb188740701\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.807176 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8xmr\" (UniqueName: \"kubernetes.io/projected/b1a72d2d-9728-492f-9e18-0cb188740701-kube-api-access-f8xmr\") pod \"b1a72d2d-9728-492f-9e18-0cb188740701\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.807322 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-utilities\") pod \"b1a72d2d-9728-492f-9e18-0cb188740701\" (UID: \"b1a72d2d-9728-492f-9e18-0cb188740701\") " Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.808041 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-utilities" (OuterVolumeSpecName: "utilities") pod "b1a72d2d-9728-492f-9e18-0cb188740701" (UID: "b1a72d2d-9728-492f-9e18-0cb188740701"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.819849 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a72d2d-9728-492f-9e18-0cb188740701-kube-api-access-f8xmr" (OuterVolumeSpecName: "kube-api-access-f8xmr") pod "b1a72d2d-9728-492f-9e18-0cb188740701" (UID: "b1a72d2d-9728-492f-9e18-0cb188740701"). InnerVolumeSpecName "kube-api-access-f8xmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.832329 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b1a72d2d-9728-492f-9e18-0cb188740701" (UID: "b1a72d2d-9728-492f-9e18-0cb188740701"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.909558 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.909598 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b1a72d2d-9728-492f-9e18-0cb188740701-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:53 crc kubenswrapper[4979]: I1204 12:07:53.909613 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8xmr\" (UniqueName: \"kubernetes.io/projected/b1a72d2d-9728-492f-9e18-0cb188740701-kube-api-access-f8xmr\") on node \"crc\" DevicePath \"\"" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.344348 4979 generic.go:334] "Generic (PLEG): container finished" podID="b1a72d2d-9728-492f-9e18-0cb188740701" containerID="ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a" exitCode=0 Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.344391 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmrqp" event={"ID":"b1a72d2d-9728-492f-9e18-0cb188740701","Type":"ContainerDied","Data":"ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a"} Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.344673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nmrqp" event={"ID":"b1a72d2d-9728-492f-9e18-0cb188740701","Type":"ContainerDied","Data":"6db0ed631a70e7656f2a6aed945b1cffe5305574a0bbd9f7cb8080934c48db6a"} Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.344694 4979 scope.go:117] "RemoveContainer" containerID="ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.344485 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nmrqp" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.378001 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmrqp"] Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.386836 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nmrqp"] Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.387842 4979 scope.go:117] "RemoveContainer" containerID="240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.424193 4979 scope.go:117] "RemoveContainer" containerID="166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.448208 4979 scope.go:117] "RemoveContainer" containerID="ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a" Dec 04 12:07:54 crc kubenswrapper[4979]: E1204 12:07:54.448728 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a\": container with ID starting with ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a not found: ID does not exist" containerID="ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.448783 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a"} err="failed to get container status \"ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a\": rpc error: code = NotFound desc = could not find container \"ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a\": container with ID starting with ef6652602bb3300d626009ba0a0c298fc61a05f777b9fbea759d1fa2cebc706a not found: ID does not exist" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.448818 4979 scope.go:117] "RemoveContainer" containerID="240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098" Dec 04 12:07:54 crc kubenswrapper[4979]: E1204 12:07:54.451144 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098\": container with ID starting with 240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098 not found: ID does not exist" containerID="240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.451233 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098"} err="failed to get container status \"240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098\": rpc error: code = NotFound desc = could not find container \"240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098\": container with ID starting with 240712e09505ad2b8adc73f895fa904fb3a465cd3afddc4c2fdbca70d046d098 not found: ID does not exist" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.451648 4979 scope.go:117] "RemoveContainer" containerID="166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f" Dec 04 12:07:54 crc kubenswrapper[4979]: E1204 12:07:54.452024 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f\": container with ID starting with 166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f not found: ID does not exist" containerID="166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f" Dec 04 12:07:54 crc kubenswrapper[4979]: I1204 12:07:54.452076 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f"} err="failed to get container status \"166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f\": rpc error: code = NotFound desc = could not find container \"166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f\": container with ID starting with 166e37a4889b7172d40101517351dc3eef1b45e3889ae5a4b7be3bb2d288c46f not found: ID does not exist" Dec 04 12:07:56 crc kubenswrapper[4979]: I1204 12:07:56.211407 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" path="/var/lib/kubelet/pods/b1a72d2d-9728-492f-9e18-0cb188740701/volumes" Dec 04 12:07:58 crc kubenswrapper[4979]: I1204 12:07:58.040705 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:07:58 crc kubenswrapper[4979]: I1204 12:07:58.041106 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.349855 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.488718 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance2ebe-account-delete-wzsfb"] Dec 04 12:08:06 crc kubenswrapper[4979]: E1204 12:08:06.489464 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="registry-server" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.489482 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="registry-server" Dec 04 12:08:06 crc kubenswrapper[4979]: E1204 12:08:06.489502 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="extract-utilities" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.489508 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="extract-utilities" Dec 04 12:08:06 crc kubenswrapper[4979]: E1204 12:08:06.489524 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="extract-content" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.489530 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="extract-content" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.489711 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a72d2d-9728-492f-9e18-0cb188740701" containerName="registry-server" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.490645 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.534369 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance2ebe-account-delete-wzsfb"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.559392 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.559639 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="e8ecd6fb-35b6-4b0d-b236-71390feeed34" containerName="openstackclient" containerID="cri-o://7c53f35962685fc14e259872b8140e5617d6f353e0dc5a20a4f05710bcea99f1" gracePeriod=2 Dec 04 12:08:06 crc kubenswrapper[4979]: E1204 12:08:06.561046 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 12:08:06 crc kubenswrapper[4979]: E1204 12:08:06.561094 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data podName:0e592d2c-bb50-44b9-bd85-fdf94ee1ac48 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:07.061076151 +0000 UTC m=+1511.335371965 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data") pod "rabbitmq-server-0" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48") : configmap "rabbitmq-config-data" not found Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.581962 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.582725 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="openstack-network-exporter" containerID="cri-o://da2072151fb1ba04ce685c63911129fe34aff9fc376cba5805e8cd1d4a2e02b6" gracePeriod=300 Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.588855 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.661404 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06a4c5f-10af-4f68-a303-f9342b97172b-operator-scripts\") pod \"glance2ebe-account-delete-wzsfb\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.661545 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2w9k4\" (UniqueName: \"kubernetes.io/projected/b06a4c5f-10af-4f68-a303-f9342b97172b-kube-api-access-2w9k4\") pod \"glance2ebe-account-delete-wzsfb\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.695527 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement375f-account-delete-28b5l"] Dec 04 12:08:06 crc kubenswrapper[4979]: E1204 12:08:06.696589 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8ecd6fb-35b6-4b0d-b236-71390feeed34" containerName="openstackclient" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.696612 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8ecd6fb-35b6-4b0d-b236-71390feeed34" containerName="openstackclient" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.696996 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8ecd6fb-35b6-4b0d-b236-71390feeed34" containerName="openstackclient" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.698314 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.795253 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec43b717-4a9b-47b0-9007-9589c7895729-operator-scripts\") pod \"placement375f-account-delete-28b5l\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.795399 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2w9k4\" (UniqueName: \"kubernetes.io/projected/b06a4c5f-10af-4f68-a303-f9342b97172b-kube-api-access-2w9k4\") pod \"glance2ebe-account-delete-wzsfb\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.795525 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb8d9\" (UniqueName: \"kubernetes.io/projected/ec43b717-4a9b-47b0-9007-9589c7895729-kube-api-access-lb8d9\") pod \"placement375f-account-delete-28b5l\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.795586 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06a4c5f-10af-4f68-a303-f9342b97172b-operator-scripts\") pod \"glance2ebe-account-delete-wzsfb\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.796382 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06a4c5f-10af-4f68-a303-f9342b97172b-operator-scripts\") pod \"glance2ebe-account-delete-wzsfb\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.815823 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement375f-account-delete-28b5l"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.867993 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2w9k4\" (UniqueName: \"kubernetes.io/projected/b06a4c5f-10af-4f68-a303-f9342b97172b-kube-api-access-2w9k4\") pod \"glance2ebe-account-delete-wzsfb\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.880134 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-sb-0" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="ovsdbserver-sb" containerID="cri-o://0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac" gracePeriod=300 Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.881897 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-8kdt5"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.892620 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-8kdt5"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.904777 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb8d9\" (UniqueName: \"kubernetes.io/projected/ec43b717-4a9b-47b0-9007-9589c7895729-kube-api-access-lb8d9\") pod \"placement375f-account-delete-28b5l\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.904928 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec43b717-4a9b-47b0-9007-9589c7895729-operator-scripts\") pod \"placement375f-account-delete-28b5l\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.906203 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec43b717-4a9b-47b0-9007-9589c7895729-operator-scripts\") pod \"placement375f-account-delete-28b5l\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.920216 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.933654 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.933878 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="ovn-northd" containerID="cri-o://4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" gracePeriod=30 Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.934219 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-northd-0" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="openstack-network-exporter" containerID="cri-o://104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024" gracePeriod=30 Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.937319 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-s2g8b"] Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.939478 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb8d9\" (UniqueName: \"kubernetes.io/projected/ec43b717-4a9b-47b0-9007-9589c7895729-kube-api-access-lb8d9\") pod \"placement375f-account-delete-28b5l\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:06 crc kubenswrapper[4979]: I1204 12:08:06.964451 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-s2g8b"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.000374 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x2hnf"] Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.009411 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.009466 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data podName:0fad7195-d19a-48ce-ad3d-4c67b53d2974 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:07.509451642 +0000 UTC m=+1511.783747446 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data") pod "rabbitmq-cell1-server-0" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974") : configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.021864 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder37f5-account-delete-4wn9x"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.022977 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.031551 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder37f5-account-delete-4wn9x"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.058550 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-v6652"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.058842 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-metrics-v6652" podUID="70791582-cb57-4564-8a5b-9953f4f22c23" containerName="openstack-network-exporter" containerID="cri-o://992bc26c538b977fe854be958e44ced58bdcb0d9b0705720d5e5a3f11286c1a4" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.103616 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.110463 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzcr9\" (UniqueName: \"kubernetes.io/projected/1341b825-bf58-4a12-af3e-40411a861e8f-kube-api-access-jzcr9\") pod \"cinder37f5-account-delete-4wn9x\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.110567 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts\") pod \"cinder37f5-account-delete-4wn9x\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.110723 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.110773 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data podName:0e592d2c-bb50-44b9-bd85-fdf94ee1ac48 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:08.11076024 +0000 UTC m=+1512.385056044 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data") pod "rabbitmq-server-0" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48") : configmap "rabbitmq-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.116518 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-x54hj"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.137531 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.169343 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-scv9t"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.201622 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-scv9t"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.214381 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzcr9\" (UniqueName: \"kubernetes.io/projected/1341b825-bf58-4a12-af3e-40411a861e8f-kube-api-access-jzcr9\") pod \"cinder37f5-account-delete-4wn9x\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.214469 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts\") pod \"cinder37f5-account-delete-4wn9x\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.215180 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts\") pod \"cinder37f5-account-delete-4wn9x\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.222357 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron47fe-account-delete-kqc4q"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.226845 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.243320 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-2465j"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.253848 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-2465j"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.273023 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzcr9\" (UniqueName: \"kubernetes.io/projected/1341b825-bf58-4a12-af3e-40411a861e8f-kube-api-access-jzcr9\") pod \"cinder37f5-account-delete-4wn9x\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.274739 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac is running failed: container process not found" containerID="0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.283129 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron47fe-account-delete-kqc4q"] Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.297208 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac is running failed: container process not found" containerID="0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.312852 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac is running failed: container process not found" containerID="0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.312918 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-sb-0" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="ovsdbserver-sb" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.313481 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbicanee53-account-delete-gpkgc"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.314640 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.316105 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csgn7\" (UniqueName: \"kubernetes.io/projected/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-kube-api-access-csgn7\") pod \"neutron47fe-account-delete-kqc4q\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.316135 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-operator-scripts\") pod \"neutron47fe-account-delete-kqc4q\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.316744 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.316955 4979 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.317012 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:07.816990082 +0000 UTC m=+1512.091285946 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.317255 4979 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.317279 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:07.81727196 +0000 UTC m=+1512.091567754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scheduler-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.349500 4979 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.349597 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:07.849575979 +0000 UTC m=+1512.123871783 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scripts" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.366603 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanee53-account-delete-gpkgc"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.418663 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5hnp\" (UniqueName: \"kubernetes.io/projected/a889a605-1dbe-4a0c-a91a-3a9d2153543a-kube-api-access-n5hnp\") pod \"barbicanee53-account-delete-gpkgc\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.419042 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts\") pod \"barbicanee53-account-delete-gpkgc\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.419151 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csgn7\" (UniqueName: \"kubernetes.io/projected/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-kube-api-access-csgn7\") pod \"neutron47fe-account-delete-kqc4q\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.419191 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-operator-scripts\") pod \"neutron47fe-account-delete-kqc4q\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.420439 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-operator-scripts\") pod \"neutron47fe-account-delete-kqc4q\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.451373 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xmtnh"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.470170 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csgn7\" (UniqueName: \"kubernetes.io/projected/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-kube-api-access-csgn7\") pod \"neutron47fe-account-delete-kqc4q\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.526454 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5hnp\" (UniqueName: \"kubernetes.io/projected/a889a605-1dbe-4a0c-a91a-3a9d2153543a-kube-api-access-n5hnp\") pod \"barbicanee53-account-delete-gpkgc\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.526495 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts\") pod \"barbicanee53-account-delete-gpkgc\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.526758 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.526800 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data podName:0fad7195-d19a-48ce-ad3d-4c67b53d2974 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:08.52678739 +0000 UTC m=+1512.801083194 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data") pod "rabbitmq-cell1-server-0" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974") : configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.527074 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xmtnh"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.527348 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts\") pod \"barbicanee53-account-delete-gpkgc\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.557100 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5hnp\" (UniqueName: \"kubernetes.io/projected/a889a605-1dbe-4a0c-a91a-3a9d2153543a-kube-api-access-n5hnp\") pod \"barbicanee53-account-delete-gpkgc\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.605337 4979 generic.go:334] "Generic (PLEG): container finished" podID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerID="104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024" exitCode=2 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.605774 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novacell02b03-account-delete-hf9n8"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.607414 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d5e30fc1-acc8-448a-a9e8-490bda22e24e","Type":"ContainerDied","Data":"104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024"} Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.607503 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.622161 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell02b03-account-delete-hf9n8"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.643791 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bdd8e97d-9bc0-4e10-984b-7d92698e8bd6/ovsdbserver-sb/0.log" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.643833 4979 generic.go:334] "Generic (PLEG): container finished" podID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerID="da2072151fb1ba04ce685c63911129fe34aff9fc376cba5805e8cd1d4a2e02b6" exitCode=2 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.643856 4979 generic.go:334] "Generic (PLEG): container finished" podID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerID="0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac" exitCode=143 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.643912 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6","Type":"ContainerDied","Data":"da2072151fb1ba04ce685c63911129fe34aff9fc376cba5805e8cd1d4a2e02b6"} Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.643938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6","Type":"ContainerDied","Data":"0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac"} Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.646889 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/novaapic904-account-delete-cvwnr"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.648092 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.658148 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.681990 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic904-account-delete-cvwnr"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.686924 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v6652_70791582-cb57-4564-8a5b-9953f4f22c23/openstack-network-exporter/0.log" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.689059 4979 generic.go:334] "Generic (PLEG): container finished" podID="70791582-cb57-4564-8a5b-9953f4f22c23" containerID="992bc26c538b977fe854be958e44ced58bdcb0d9b0705720d5e5a3f11286c1a4" exitCode=2 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.689107 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v6652" event={"ID":"70791582-cb57-4564-8a5b-9953f4f22c23","Type":"ContainerDied","Data":"992bc26c538b977fe854be958e44ced58bdcb0d9b0705720d5e5a3f11286c1a4"} Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.700878 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5667db5f8d-q4ffn"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.701104 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5667db5f8d-q4ffn" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-log" containerID="cri-o://89858cce1ad717d7a0855a8184b9e88c16893052de81ca4a07efe6a69e691109" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.701200 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-5667db5f8d-q4ffn" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-api" containerID="cri-o://dfee452098fffdbca65d5be95466eeb6448db8377ba02e117a1cd9cea7962b7f" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.713339 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.739497 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts\") pod \"novaapic904-account-delete-cvwnr\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.739584 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts\") pod \"novacell02b03-account-delete-hf9n8\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.744710 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xwsw\" (UniqueName: \"kubernetes.io/projected/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-kube-api-access-9xwsw\") pod \"novacell02b03-account-delete-hf9n8\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.752997 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fh88\" (UniqueName: \"kubernetes.io/projected/f9b386d0-b110-48a9-b6fe-377f0faeedc0-kube-api-access-2fh88\") pod \"novaapic904-account-delete-cvwnr\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.754234 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bnbw6"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.755042 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerName="dnsmasq-dns" containerID="cri-o://afd5a98b65f1b147b6118a4045a232bfd6c4bd85bfd36e787d65bf8b8b8468a2" gracePeriod=10 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.774154 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.774762 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="openstack-network-exporter" containerID="cri-o://43237a8cc1d10bdde09a5a677588338264c1215b57330e0176e9bd4306ed5bf6" gracePeriod=300 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.820592 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.820953 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-log" containerID="cri-o://0b18da01797ecd4c68f4b61b2e504256585316a4b410185ba40df8298d3b3107" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.839079 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-httpd" containerID="cri-o://02024a485457048e2f54ce565223a36a682c9da052f1a487dff4004a22a382ee" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.856893 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts\") pod \"novacell02b03-account-delete-hf9n8\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.856957 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xwsw\" (UniqueName: \"kubernetes.io/projected/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-kube-api-access-9xwsw\") pod \"novacell02b03-account-delete-hf9n8\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.857075 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fh88\" (UniqueName: \"kubernetes.io/projected/f9b386d0-b110-48a9-b6fe-377f0faeedc0-kube-api-access-2fh88\") pod \"novaapic904-account-delete-cvwnr\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.857216 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts\") pod \"novaapic904-account-delete-cvwnr\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.857699 4979 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.857777 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:08.8577551 +0000 UTC m=+1513.132050904 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.858144 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts\") pod \"novaapic904-account-delete-cvwnr\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.858245 4979 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.861441 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts\") pod \"novacell02b03-account-delete-hf9n8\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.862106 4979 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.885933 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.886637 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-log" containerID="cri-o://7341c4702a0b8160bf9a2a2ae8f818d6b9089d0a90f6a4ed29d2510863a765d0" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.886777 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-httpd" containerID="cri-o://063e5850e136541b04a3fbcd523818e0db023f1015cc97b6258d7d2440f0fc84" gracePeriod=30 Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.887766 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:08.858281685 +0000 UTC m=+1513.132577489 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scripts" not found Dec 04 12:08:07 crc kubenswrapper[4979]: E1204 12:08:07.887854 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:08.8878183 +0000 UTC m=+1513.162114104 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scheduler-config-data" not found Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.913084 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xwsw\" (UniqueName: \"kubernetes.io/projected/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-kube-api-access-9xwsw\") pod \"novacell02b03-account-delete-hf9n8\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.923710 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fh88\" (UniqueName: \"kubernetes.io/projected/f9b386d0-b110-48a9-b6fe-377f0faeedc0-kube-api-access-2fh88\") pod \"novaapic904-account-delete-cvwnr\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.946281 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-t2bg8"] Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.974555 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:07 crc kubenswrapper[4979]: I1204 12:08:07.987930 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-t2bg8"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.000837 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovsdbserver-nb-0" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="ovsdbserver-nb" containerID="cri-o://c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef" gracePeriod=300 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.005792 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.006662 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-server" containerID="cri-o://69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.007930 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="swift-recon-cron" containerID="cri-o://a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008004 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="rsync" containerID="cri-o://1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008037 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-expirer" containerID="cri-o://fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008068 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-updater" containerID="cri-o://edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008100 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-auditor" containerID="cri-o://fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008129 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-replicator" containerID="cri-o://58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008157 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-server" containerID="cri-o://f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008183 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-updater" containerID="cri-o://4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008211 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-auditor" containerID="cri-o://89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008238 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-replicator" containerID="cri-o://f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008266 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-server" containerID="cri-o://4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008310 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-reaper" containerID="cri-o://f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008352 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-auditor" containerID="cri-o://d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.008382 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-storage-0" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-replicator" containerID="cri-o://37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.075440 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-698495cf57-xzqk7"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.076034 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-698495cf57-xzqk7" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-api" containerID="cri-o://225feb849cbf0c27ccb08142f91bf2d2a45a52845a9b008d7dd3f641eb0c0175" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.076856 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-698495cf57-xzqk7" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-httpd" containerID="cri-o://4e4c287b0e57b02ab10e8858559ba9bd0849ef197cf2a81904670a23b3ad447d" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.179098 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.179160 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data podName:0e592d2c-bb50-44b9-bd85-fdf94ee1ac48 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:10.179147213 +0000 UTC m=+1514.453443017 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data") pod "rabbitmq-server-0" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48") : configmap "rabbitmq-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.279984 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9" path="/var/lib/kubelet/pods/257c07f0-7fa5-4ba4-9dec-e2fa0a9ab5a9/volumes" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.280859 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bc5e0d4-e60f-465c-8ed5-c9a4203f4112" path="/var/lib/kubelet/pods/8bc5e0d4-e60f-465c-8ed5-c9a4203f4112/volumes" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.297847 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0fd1874-37a5-40ab-ab1e-c18a0cffbab6" path="/var/lib/kubelet/pods/c0fd1874-37a5-40ab-ab1e-c18a0cffbab6/volumes" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.302664 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e14c1da3-5204-4d84-94fd-039b1cef8af3" path="/var/lib/kubelet/pods/e14c1da3-5204-4d84-94fd-039b1cef8af3/volumes" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.303281 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9ceee6e-4685-4dc1-8222-57cc6d51a337" path="/var/lib/kubelet/pods/e9ceee6e-4685-4dc1-8222-57cc6d51a337/volumes" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.303963 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4f85b75-ee47-4c71-822c-e30cd4f3492c" path="/var/lib/kubelet/pods/f4f85b75-ee47-4c71-822c-e30cd4f3492c/volumes" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.313976 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-lcmq5"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314287 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-lcmq5"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314321 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-cxc6d"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314339 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-cxc6d"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314353 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314378 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314629 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api-log" containerID="cri-o://07d2a54c7f114041633c7190d149a251778f6bde1cdbf1ccdf26922e6d2aeed8" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.314889 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="cinder-scheduler" containerID="cri-o://66265992f836c6a9f8e585ba56a557ed7f384463e217621e5223701c105f7441" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.315195 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api" containerID="cri-o://3d54c5d9cb36dae2f6bcd838a3e0fddaf4c2d1f19e347ee987290f673c8efd37" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.315489 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="probe" containerID="cri-o://aa4d07bf279adaae12d8aad041308e01478fd7338e4e1c8dfe3e93d2a095aa65" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.321414 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.333387 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.333713 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-log" containerID="cri-o://601adae13b36c5015c0e550950004699f70cac5db6ce0564490dc4284b7bfc50" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.333885 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-metadata" containerID="cri-o://640bdb1e4d1f6981f6d0fba2f1150c83d60fe36e91add11c47a642a8218b2507" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.344905 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.345116 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-log" containerID="cri-o://ea884876a5d807dfae591e01b960188f749527339aa8424bc250430bc801a04f" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.345260 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-api" containerID="cri-o://0d566502255055e35e3f3b677f5eefb5b5ec440a9ad5528a85688965a96d78c3" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.415599 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.475036 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-71eb-account-create-update-wdcd2"] Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.487094 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef is running failed: container process not found" containerID="c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.488261 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef is running failed: container process not found" containerID="c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.493910 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-71eb-account-create-update-wdcd2"] Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.495461 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef is running failed: container process not found" containerID="c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef" cmd=["/usr/bin/pidof","ovsdb-server"] Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.495525 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef is running failed: container process not found" probeType="Readiness" pod="openstack/ovsdbserver-nb-0" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="ovsdbserver-nb" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.529771 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-9kwrg"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.538484 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7b84847487-r8c7n"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.538709 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7b84847487-r8c7n" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker-log" containerID="cri-o://de14bd59b0b06af679f770868d3c98798bbffcb02c92b7336ab4b61c2bc90a5b" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.539095 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-worker-7b84847487-r8c7n" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker" containerID="cri-o://15976ba28c5c9feffdf2219916bb384ba02e85b8e1455eb217e64c14c2a43e76" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.555092 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-9kwrg"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.557782 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="rabbitmq" containerID="cri-o://ea56050492cab7388c6ef7c5a9645f7cda3ba4238e27b00f2097646a63874d12" gracePeriod=604800 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.592242 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57cfcdd946-gb75v"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.592777 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57cfcdd946-gb75v" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api-log" containerID="cri-o://f79822b9ecbd0ae0552d4c0bef356ac9d8b4b613b0aa7af9b7905e5627f9d542" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.592941 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-57cfcdd946-gb75v" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api" containerID="cri-o://c1697a681e440b41b665f6d4ba7c31e3a29aecbb77162bfc4b85770584ea8a4e" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.603260 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.603337 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data podName:0fad7195-d19a-48ce-ad3d-4c67b53d2974 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:10.603321063 +0000 UTC m=+1514.877616867 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data") pod "rabbitmq-cell1-server-0" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974") : configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.623703 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6bf4748b96-fzr6d"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.624342 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener-log" containerID="cri-o://4a64fd8d676b9faa863cf1c62e916da023a1db555a91558e98ac315dc5777bab" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.624836 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener" containerID="cri-o://b68d477aa81babb95a8e72c26e66dc3b949d75be0432ca0d51c1a917acf6f6e3" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.647166 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.682682 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.682912 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://676d3e7b21b50404e78713930c4c90e6f133e77d13cd221bda1604c38311a704" gracePeriod=30 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.702131 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.769699 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="rabbitmq" containerID="cri-o://a170026606f642d3fddc6953a178bf2cc7cbe7322dc226b38b4e679b7df1e228" gracePeriod=604800 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.770874 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" containerID="cri-o://1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" gracePeriod=29 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.810705 4979 generic.go:334] "Generic (PLEG): container finished" podID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerID="0b18da01797ecd4c68f4b61b2e504256585316a4b410185ba40df8298d3b3107" exitCode=143 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.810792 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d24f5a-ae53-44e3-bc4c-c56813f0f141","Type":"ContainerDied","Data":"0b18da01797ecd4c68f4b61b2e504256585316a4b410185ba40df8298d3b3107"} Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.862469 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v6652_70791582-cb57-4564-8a5b-9953f4f22c23/openstack-network-exporter/0.log" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.862549 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-v6652" event={"ID":"70791582-cb57-4564-8a5b-9953f4f22c23","Type":"ContainerDied","Data":"02d87c993f9049eaa0cb96903e59776aaa45e7d153e10fab1bbf6bb9aeef4ff1"} Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.862577 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02d87c993f9049eaa0cb96903e59776aaa45e7d153e10fab1bbf6bb9aeef4ff1" Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.932866 4979 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.932922 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:10.932908486 +0000 UTC m=+1515.207204290 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scheduler-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.932962 4979 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.932982 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:10.932976247 +0000 UTC m=+1515.207272051 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-config-data" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.933017 4979 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 04 12:08:08 crc kubenswrapper[4979]: E1204 12:08:08.933035 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:10.933028389 +0000 UTC m=+1515.207324193 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scripts" not found Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.952293 4979 generic.go:334] "Generic (PLEG): container finished" podID="c472e576-897a-4ea6-912a-2d87b2f67077" containerID="ea884876a5d807dfae591e01b960188f749527339aa8424bc250430bc801a04f" exitCode=143 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.952602 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c472e576-897a-4ea6-912a-2d87b2f67077","Type":"ContainerDied","Data":"ea884876a5d807dfae591e01b960188f749527339aa8424bc250430bc801a04f"} Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.963573 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-v6652_70791582-cb57-4564-8a5b-9953f4f22c23/openstack-network-exporter/0.log" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.963798 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.986715 4979 generic.go:334] "Generic (PLEG): container finished" podID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerID="4e4c287b0e57b02ab10e8858559ba9bd0849ef197cf2a81904670a23b3ad447d" exitCode=0 Dec 04 12:08:08 crc kubenswrapper[4979]: I1204 12:08:08.987035 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698495cf57-xzqk7" event={"ID":"d5d21f75-aea0-4306-843d-c3eb782cdd78","Type":"ContainerDied","Data":"4e4c287b0e57b02ab10e8858559ba9bd0849ef197cf2a81904670a23b3ad447d"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.054512 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"https://10.217.0.197:6080/vnc_lite.html\": dial tcp 10.217.0.197:6080: connect: connection refused" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.087906 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bdd8e97d-9bc0-4e10-984b-7d92698e8bd6/ovsdbserver-sb/0.log" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.088319 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123213 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123240 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123247 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123253 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123261 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123268 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123274 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123280 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123285 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123292 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123313 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123319 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123325 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123331 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123377 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123402 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123414 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123422 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123432 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123440 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123447 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123455 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123464 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123473 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123482 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123490 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123500 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.123512 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254"} Dec 04 12:08:09 crc kubenswrapper[4979]: E1204 12:08:09.134948 4979 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 04 12:08:09 crc kubenswrapper[4979]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 04 12:08:09 crc kubenswrapper[4979]: + source /usr/local/bin/container-scripts/functions Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNBridge=br-int Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNRemote=tcp:localhost:6642 Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNEncapType=geneve Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNAvailabilityZones= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ EnableChassisAsGateway=true Dec 04 12:08:09 crc kubenswrapper[4979]: ++ PhysicalNetworks= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNHostName= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 04 12:08:09 crc kubenswrapper[4979]: ++ ovs_dir=/var/lib/openvswitch Dec 04 12:08:09 crc kubenswrapper[4979]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 04 12:08:09 crc kubenswrapper[4979]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 04 12:08:09 crc kubenswrapper[4979]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + cleanup_ovsdb_server_semaphore Dec 04 12:08:09 crc kubenswrapper[4979]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 12:08:09 crc kubenswrapper[4979]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 04 12:08:09 crc kubenswrapper[4979]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-x54hj" message=< Dec 04 12:08:09 crc kubenswrapper[4979]: Exiting ovsdb-server (5) [ OK ] Dec 04 12:08:09 crc kubenswrapper[4979]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 04 12:08:09 crc kubenswrapper[4979]: + source /usr/local/bin/container-scripts/functions Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNBridge=br-int Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNRemote=tcp:localhost:6642 Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNEncapType=geneve Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNAvailabilityZones= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ EnableChassisAsGateway=true Dec 04 12:08:09 crc kubenswrapper[4979]: ++ PhysicalNetworks= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNHostName= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 04 12:08:09 crc kubenswrapper[4979]: ++ ovs_dir=/var/lib/openvswitch Dec 04 12:08:09 crc kubenswrapper[4979]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 04 12:08:09 crc kubenswrapper[4979]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 04 12:08:09 crc kubenswrapper[4979]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + cleanup_ovsdb_server_semaphore Dec 04 12:08:09 crc kubenswrapper[4979]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 12:08:09 crc kubenswrapper[4979]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 04 12:08:09 crc kubenswrapper[4979]: > Dec 04 12:08:09 crc kubenswrapper[4979]: E1204 12:08:09.134989 4979 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 04 12:08:09 crc kubenswrapper[4979]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Dec 04 12:08:09 crc kubenswrapper[4979]: + source /usr/local/bin/container-scripts/functions Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNBridge=br-int Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNRemote=tcp:localhost:6642 Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNEncapType=geneve Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNAvailabilityZones= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ EnableChassisAsGateway=true Dec 04 12:08:09 crc kubenswrapper[4979]: ++ PhysicalNetworks= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ OVNHostName= Dec 04 12:08:09 crc kubenswrapper[4979]: ++ DB_FILE=/etc/openvswitch/conf.db Dec 04 12:08:09 crc kubenswrapper[4979]: ++ ovs_dir=/var/lib/openvswitch Dec 04 12:08:09 crc kubenswrapper[4979]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Dec 04 12:08:09 crc kubenswrapper[4979]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Dec 04 12:08:09 crc kubenswrapper[4979]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + sleep 0.5 Dec 04 12:08:09 crc kubenswrapper[4979]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Dec 04 12:08:09 crc kubenswrapper[4979]: + cleanup_ovsdb_server_semaphore Dec 04 12:08:09 crc kubenswrapper[4979]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Dec 04 12:08:09 crc kubenswrapper[4979]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Dec 04 12:08:09 crc kubenswrapper[4979]: > pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" containerID="cri-o://64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.135035 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" containerID="cri-o://64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" gracePeriod=28 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.136287 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-combined-ca-bundle\") pod \"70791582-cb57-4564-8a5b-9953f4f22c23\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.136439 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovs-rundir\") pod \"70791582-cb57-4564-8a5b-9953f4f22c23\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.136593 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4nnm\" (UniqueName: \"kubernetes.io/projected/70791582-cb57-4564-8a5b-9953f4f22c23-kube-api-access-f4nnm\") pod \"70791582-cb57-4564-8a5b-9953f4f22c23\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.136632 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-metrics-certs-tls-certs\") pod \"70791582-cb57-4564-8a5b-9953f4f22c23\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.136693 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovn-rundir\") pod \"70791582-cb57-4564-8a5b-9953f4f22c23\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.136718 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70791582-cb57-4564-8a5b-9953f4f22c23-config\") pod \"70791582-cb57-4564-8a5b-9953f4f22c23\" (UID: \"70791582-cb57-4564-8a5b-9953f4f22c23\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.137793 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70791582-cb57-4564-8a5b-9953f4f22c23-config" (OuterVolumeSpecName: "config") pod "70791582-cb57-4564-8a5b-9953f4f22c23" (UID: "70791582-cb57-4564-8a5b-9953f4f22c23"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.140924 4979 generic.go:334] "Generic (PLEG): container finished" podID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerID="7341c4702a0b8160bf9a2a2ae8f818d6b9089d0a90f6a4ed29d2510863a765d0" exitCode=143 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.141154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c","Type":"ContainerDied","Data":"7341c4702a0b8160bf9a2a2ae8f818d6b9089d0a90f6a4ed29d2510863a765d0"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.141310 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "70791582-cb57-4564-8a5b-9953f4f22c23" (UID: "70791582-cb57-4564-8a5b-9953f4f22c23"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.144840 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovs-rundir" (OuterVolumeSpecName: "ovs-rundir") pod "70791582-cb57-4564-8a5b-9953f4f22c23" (UID: "70791582-cb57-4564-8a5b-9953f4f22c23"). InnerVolumeSpecName "ovs-rundir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.158526 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70791582-cb57-4564-8a5b-9953f4f22c23-kube-api-access-f4nnm" (OuterVolumeSpecName: "kube-api-access-f4nnm") pod "70791582-cb57-4564-8a5b-9953f4f22c23" (UID: "70791582-cb57-4564-8a5b-9953f4f22c23"). InnerVolumeSpecName "kube-api-access-f4nnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.188462 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.189141 4979 generic.go:334] "Generic (PLEG): container finished" podID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerID="601adae13b36c5015c0e550950004699f70cac5db6ce0564490dc4284b7bfc50" exitCode=143 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.189206 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3238d4e5-e968-4d66-87d6-7397f48e9714","Type":"ContainerDied","Data":"601adae13b36c5015c0e550950004699f70cac5db6ce0564490dc4284b7bfc50"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.204492 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70791582-cb57-4564-8a5b-9953f4f22c23" (UID: "70791582-cb57-4564-8a5b-9953f4f22c23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.214333 4979 generic.go:334] "Generic (PLEG): container finished" podID="e8ecd6fb-35b6-4b0d-b236-71390feeed34" containerID="7c53f35962685fc14e259872b8140e5617d6f353e0dc5a20a4f05710bcea99f1" exitCode=137 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.226749 4979 generic.go:334] "Generic (PLEG): container finished" podID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerID="de14bd59b0b06af679f770868d3c98798bbffcb02c92b7336ab4b61c2bc90a5b" exitCode=143 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.226858 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b84847487-r8c7n" event={"ID":"869a1863-ce1d-4acd-9d50-6987b60e05f0","Type":"ContainerDied","Data":"de14bd59b0b06af679f770868d3c98798bbffcb02c92b7336ab4b61c2bc90a5b"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.233194 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerName="galera" containerID="cri-o://fe043f249efb9dfd8bb0e3a20f1360262cf5b2c026e1870681a1fabdae2c45d6" gracePeriod=30 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241177 4979 generic.go:334] "Generic (PLEG): container finished" podID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerID="89858cce1ad717d7a0855a8184b9e88c16893052de81ca4a07efe6a69e691109" exitCode=143 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241191 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-sb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241409 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdb-rundir\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241433 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5667db5f8d-q4ffn" event={"ID":"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd","Type":"ContainerDied","Data":"89858cce1ad717d7a0855a8184b9e88c16893052de81ca4a07efe6a69e691109"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241466 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qrt2\" (UniqueName: \"kubernetes.io/projected/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-kube-api-access-8qrt2\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241530 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-combined-ca-bundle\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241579 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-config\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241662 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-scripts\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.242932 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdbserver-sb-tls-certs\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.242959 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-metrics-certs-tls-certs\") pod \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\" (UID: \"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.243773 4979 reconciler_common.go:293] "Volume detached for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovs-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.243793 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4nnm\" (UniqueName: \"kubernetes.io/projected/70791582-cb57-4564-8a5b-9953f4f22c23-kube-api-access-f4nnm\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.243804 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/70791582-cb57-4564-8a5b-9953f4f22c23-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.243813 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70791582-cb57-4564-8a5b-9953f4f22c23-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.244077 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.241776 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.242485 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-scripts" (OuterVolumeSpecName: "scripts") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.242851 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-config" (OuterVolumeSpecName: "config") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258435 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "ovndbcluster-sb-etc-ovn") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258549 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-kube-api-access-8qrt2" (OuterVolumeSpecName: "kube-api-access-8qrt2") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "kube-api-access-8qrt2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258754 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ef961b3d-b927-42e5-91b9-eb966bdee861/ovsdbserver-nb/0.log" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258802 4979 generic.go:334] "Generic (PLEG): container finished" podID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerID="43237a8cc1d10bdde09a5a677588338264c1215b57330e0176e9bd4306ed5bf6" exitCode=2 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258818 4979 generic.go:334] "Generic (PLEG): container finished" podID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerID="c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef" exitCode=143 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258888 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ef961b3d-b927-42e5-91b9-eb966bdee861","Type":"ContainerDied","Data":"43237a8cc1d10bdde09a5a677588338264c1215b57330e0176e9bd4306ed5bf6"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.258919 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ef961b3d-b927-42e5-91b9-eb966bdee861","Type":"ContainerDied","Data":"c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.280511 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "70791582-cb57-4564-8a5b-9953f4f22c23" (UID: "70791582-cb57-4564-8a5b-9953f4f22c23"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.280892 4979 generic.go:334] "Generic (PLEG): container finished" podID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerID="07d2a54c7f114041633c7190d149a251778f6bde1cdbf1ccdf26922e6d2aeed8" exitCode=143 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.281013 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4ad703ff-c521-40f1-9fa4-09fca7726079","Type":"ContainerDied","Data":"07d2a54c7f114041633c7190d149a251778f6bde1cdbf1ccdf26922e6d2aeed8"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.285252 4979 generic.go:334] "Generic (PLEG): container finished" podID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerID="afd5a98b65f1b147b6118a4045a232bfd6c4bd85bfd36e787d65bf8b8b8468a2" exitCode=0 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.285316 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" event={"ID":"3015ecd7-aacc-41ac-897f-083c4cb5bd62","Type":"ContainerDied","Data":"afd5a98b65f1b147b6118a4045a232bfd6c4bd85bfd36e787d65bf8b8b8468a2"} Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.285350 4979 scope.go:117] "RemoveContainer" containerID="afd5a98b65f1b147b6118a4045a232bfd6c4bd85bfd36e787d65bf8b8b8468a2" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.285470 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-89c5cd4d5-bnbw6" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.310400 4979 scope.go:117] "RemoveContainer" containerID="cd301899a7d93d2b87acbfde23d0f5ef3985c8227fa2a11b27c9b5365050435f" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.348178 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-swift-storage-0\") pod \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.348238 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-svc\") pod \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.348392 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-config\") pod \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.348502 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-nb\") pod \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.348529 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-sb\") pod \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.348561 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4rpr\" (UniqueName: \"kubernetes.io/projected/3015ecd7-aacc-41ac-897f-083c4cb5bd62-kube-api-access-m4rpr\") pod \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\" (UID: \"3015ecd7-aacc-41ac-897f-083c4cb5bd62\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.349025 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qrt2\" (UniqueName: \"kubernetes.io/projected/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-kube-api-access-8qrt2\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.349041 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.349050 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.349070 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.349079 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.349087 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/70791582-cb57-4564-8a5b-9953f4f22c23-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.352709 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.385794 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3015ecd7-aacc-41ac-897f-083c4cb5bd62-kube-api-access-m4rpr" (OuterVolumeSpecName: "kube-api-access-m4rpr") pod "3015ecd7-aacc-41ac-897f-083c4cb5bd62" (UID: "3015ecd7-aacc-41ac-897f-083c4cb5bd62"). InnerVolumeSpecName "kube-api-access-m4rpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.406591 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-26lwp"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.423349 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.424163 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.424567 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerName="nova-cell1-conductor-conductor" containerID="cri-o://8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" gracePeriod=30 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.451735 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdbserver-sb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-sb-tls-certs") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "ovsdbserver-sb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.458378 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-ovsdbserver-sb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.458414 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.458426 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.458444 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4rpr\" (UniqueName: \"kubernetes.io/projected/3015ecd7-aacc-41ac-897f-083c4cb5bd62-kube-api-access-m4rpr\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.460610 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-26lwp"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.473399 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-config" (OuterVolumeSpecName: "config") pod "3015ecd7-aacc-41ac-897f-083c4cb5bd62" (UID: "3015ecd7-aacc-41ac-897f-083c4cb5bd62"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.527829 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3015ecd7-aacc-41ac-897f-083c4cb5bd62" (UID: "3015ecd7-aacc-41ac-897f-083c4cb5bd62"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: E1204 12:08:09.532795 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.533399 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3015ecd7-aacc-41ac-897f-083c4cb5bd62" (UID: "3015ecd7-aacc-41ac-897f-083c4cb5bd62"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: E1204 12:08:09.537124 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:09 crc kubenswrapper[4979]: E1204 12:08:09.554460 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:09 crc kubenswrapper[4979]: E1204 12:08:09.554531 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerName="nova-cell1-conductor-conductor" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.556333 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3015ecd7-aacc-41ac-897f-083c4cb5bd62" (UID: "3015ecd7-aacc-41ac-897f-083c4cb5bd62"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.556414 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3015ecd7-aacc-41ac-897f-083c4cb5bd62" (UID: "3015ecd7-aacc-41ac-897f-083c4cb5bd62"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.574354 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.588252 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k59t5"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.590085 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.591546 4979 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.591875 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.591961 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.592030 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3015ecd7-aacc-41ac-897f-083c4cb5bd62-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.604029 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.604642 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" containerName="nova-cell0-conductor-conductor" containerID="cri-o://fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca" gracePeriod=30 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.616618 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-k59t5"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.618453 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" (UID: "bdd8e97d-9bc0-4e10-984b-7d92698e8bd6"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.632680 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.632923 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="77ca261e-57dc-4b57-a80b-e7def9b113d5" containerName="nova-scheduler-scheduler" containerID="cri-o://f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" gracePeriod=30 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.643923 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance2ebe-account-delete-wzsfb"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.652231 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement375f-account-delete-28b5l"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.662077 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bnbw6"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.672899 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-89c5cd4d5-bnbw6"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.692996 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dn9rq\" (UniqueName: \"kubernetes.io/projected/e8ecd6fb-35b6-4b0d-b236-71390feeed34-kube-api-access-dn9rq\") pod \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.693095 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-combined-ca-bundle\") pod \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.693249 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config\") pod \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.693267 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config-secret\") pod \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\" (UID: \"e8ecd6fb-35b6-4b0d-b236-71390feeed34\") " Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.693731 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.717097 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8ecd6fb-35b6-4b0d-b236-71390feeed34-kube-api-access-dn9rq" (OuterVolumeSpecName: "kube-api-access-dn9rq") pod "e8ecd6fb-35b6-4b0d-b236-71390feeed34" (UID: "e8ecd6fb-35b6-4b0d-b236-71390feeed34"). InnerVolumeSpecName "kube-api-access-dn9rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.744462 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e8ecd6fb-35b6-4b0d-b236-71390feeed34" (UID: "e8ecd6fb-35b6-4b0d-b236-71390feeed34"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.769770 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e8ecd6fb-35b6-4b0d-b236-71390feeed34" (UID: "e8ecd6fb-35b6-4b0d-b236-71390feeed34"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.770524 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8ecd6fb-35b6-4b0d-b236-71390feeed34" (UID: "e8ecd6fb-35b6-4b0d-b236-71390feeed34"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.800596 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.800626 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.800636 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dn9rq\" (UniqueName: \"kubernetes.io/projected/e8ecd6fb-35b6-4b0d-b236-71390feeed34-kube-api-access-dn9rq\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.800644 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8ecd6fb-35b6-4b0d-b236-71390feeed34-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.850193 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron47fe-account-delete-kqc4q"] Dec 04 12:08:09 crc kubenswrapper[4979]: W1204 12:08:09.864653 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b2ecb0e_a2d5_482f_8f10_3b32a0a3d941.slice/crio-3a7a77cfb23ce2082033ee114f5048e53209c125345774e7c7842daafb4fddd4 WatchSource:0}: Error finding container 3a7a77cfb23ce2082033ee114f5048e53209c125345774e7c7842daafb4fddd4: Status 404 returned error can't find the container with id 3a7a77cfb23ce2082033ee114f5048e53209c125345774e7c7842daafb4fddd4 Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.965884 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novaapic904-account-delete-cvwnr"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.976009 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/novacell02b03-account-delete-hf9n8"] Dec 04 12:08:09 crc kubenswrapper[4979]: I1204 12:08:09.990861 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbicanee53-account-delete-gpkgc"] Dec 04 12:08:10 crc kubenswrapper[4979]: W1204 12:08:10.034457 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1341b825_bf58_4a12_af3e_40411a861e8f.slice/crio-ae2aa6d8b077f554ae691f32a9697c1526209e0172b29c78569ae88e8a8529ba WatchSource:0}: Error finding container ae2aa6d8b077f554ae691f32a9697c1526209e0172b29c78569ae88e8a8529ba: Status 404 returned error can't find the container with id ae2aa6d8b077f554ae691f32a9697c1526209e0172b29c78569ae88e8a8529ba Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.063461 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder37f5-account-delete-4wn9x"] Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.108763 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ef961b3d-b927-42e5-91b9-eb966bdee861/ovsdbserver-nb/0.log" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.108899 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.222911 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-metrics-certs-tls-certs\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.222966 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdb-rundir\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.222999 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-combined-ca-bundle\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.223030 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkp69\" (UniqueName: \"kubernetes.io/projected/ef961b3d-b927-42e5-91b9-eb966bdee861-kube-api-access-gkp69\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.223148 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdbserver-nb-tls-certs\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.223191 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-scripts\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.223242 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-config\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.223262 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndbcluster-nb-etc-ovn\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ef961b3d-b927-42e5-91b9-eb966bdee861\" (UID: \"ef961b3d-b927-42e5-91b9-eb966bdee861\") " Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.223789 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.223838 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data podName:0e592d2c-bb50-44b9-bd85-fdf94ee1ac48 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.22382307 +0000 UTC m=+1518.498118874 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data") pod "rabbitmq-server-0" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48") : configmap "rabbitmq-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.226725 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-config" (OuterVolumeSpecName: "config") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.227127 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-scripts" (OuterVolumeSpecName: "scripts") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.230599 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdb-rundir" (OuterVolumeSpecName: "ovsdb-rundir") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "ovsdb-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.240578 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0175c40e-2589-44d3-adec-95479286dedc" path="/var/lib/kubelet/pods/0175c40e-2589-44d3-adec-95479286dedc/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.245648 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="113cba3a-dbbb-44a6-8342-3e287e77cb4d" path="/var/lib/kubelet/pods/113cba3a-dbbb-44a6-8342-3e287e77cb4d/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.246340 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" path="/var/lib/kubelet/pods/3015ecd7-aacc-41ac-897f-083c4cb5bd62/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.246471 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "ovndbcluster-nb-etc-ovn") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.246968 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36c9067d-e2c5-443d-922a-78441d66f69b" path="/var/lib/kubelet/pods/36c9067d-e2c5-443d-922a-78441d66f69b/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.255157 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b62e2c4-065e-48b2-abc0-03f445f8d787" path="/var/lib/kubelet/pods/7b62e2c4-065e-48b2-abc0-03f445f8d787/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.255824 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="814e268c-44eb-4a89-a58f-a38ac92fde57" path="/var/lib/kubelet/pods/814e268c-44eb-4a89-a58f-a38ac92fde57/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.256487 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5529cf7-2998-44bc-bb8f-3cc8da54450d" path="/var/lib/kubelet/pods/c5529cf7-2998-44bc-bb8f-3cc8da54450d/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.257654 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8ecd6fb-35b6-4b0d-b236-71390feeed34" path="/var/lib/kubelet/pods/e8ecd6fb-35b6-4b0d-b236-71390feeed34/volumes" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.258481 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef961b3d-b927-42e5-91b9-eb966bdee861-kube-api-access-gkp69" (OuterVolumeSpecName: "kube-api-access-gkp69") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "kube-api-access-gkp69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.326577 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdb-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.326604 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkp69\" (UniqueName: \"kubernetes.io/projected/ef961b3d-b927-42e5-91b9-eb966bdee861-kube-api-access-gkp69\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.326614 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.326622 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef961b3d-b927-42e5-91b9-eb966bdee861-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.326641 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.356731 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.366689 4979 generic.go:334] "Generic (PLEG): container finished" podID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerID="f79822b9ecbd0ae0552d4c0bef356ac9d8b4b613b0aa7af9b7905e5627f9d542" exitCode=143 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.369581 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.378332 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.379438 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdbserver-nb-tls-certs" (OuterVolumeSpecName: "ovsdbserver-nb-tls-certs") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "ovsdbserver-nb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.379520 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_ef961b3d-b927-42e5-91b9-eb966bdee861/ovsdbserver-nb/0.log" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.379697 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.379805 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cfcdd946-gb75v" event={"ID":"1ba50945-6ce9-4a71-9f58-e436b646c779","Type":"ContainerDied","Data":"f79822b9ecbd0ae0552d4c0bef356ac9d8b4b613b0aa7af9b7905e5627f9d542"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.379856 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"ef961b3d-b927-42e5-91b9-eb966bdee861","Type":"ContainerDied","Data":"2f24c9ff025d883e5b4288cd6adff9b63f9decd968e600453fdad6a9eb5c359b"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.379884 4979 scope.go:117] "RemoveContainer" containerID="7c53f35962685fc14e259872b8140e5617d6f353e0dc5a20a4f05710bcea99f1" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.383122 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance2ebe-account-delete-wzsfb" event={"ID":"b06a4c5f-10af-4f68-a303-f9342b97172b","Type":"ContainerStarted","Data":"a7f3cfcc6ea9f0d1247475acda7e385c450f0f81f09b257c0d8728723a8cc66c"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.389524 4979 generic.go:334] "Generic (PLEG): container finished" podID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerID="4a64fd8d676b9faa863cf1c62e916da023a1db555a91558e98ac315dc5777bab" exitCode=143 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.389614 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" event={"ID":"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a","Type":"ContainerDied","Data":"4a64fd8d676b9faa863cf1c62e916da023a1db555a91558e98ac315dc5777bab"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.412666 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_bdd8e97d-9bc0-4e10-984b-7d92698e8bd6/ovsdbserver-sb/0.log" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.412764 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"bdd8e97d-9bc0-4e10-984b-7d92698e8bd6","Type":"ContainerDied","Data":"566bb3521fcc918076339895b00651591a9d329423b5a31229a6b38c106e9c2d"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.412927 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.432330 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.432536 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-ovsdbserver-nb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.432548 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.433799 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder37f5-account-delete-4wn9x" event={"ID":"1341b825-bf58-4a12-af3e-40411a861e8f","Type":"ContainerStarted","Data":"ae2aa6d8b077f554ae691f32a9697c1526209e0172b29c78569ae88e8a8529ba"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.454471 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02b03-account-delete-hf9n8" event={"ID":"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4","Type":"ContainerStarted","Data":"bdb561f2fd816a098e8a480f61594120050d3c8b7811b4f236cc21ce18301d04"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.456669 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron47fe-account-delete-kqc4q" event={"ID":"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941","Type":"ContainerStarted","Data":"3a7a77cfb23ce2082033ee114f5048e53209c125345774e7c7842daafb4fddd4"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.457654 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic904-account-delete-cvwnr" event={"ID":"f9b386d0-b110-48a9-b6fe-377f0faeedc0","Type":"ContainerStarted","Data":"8caf959e5d5213b30979b37f6649552a9ef3dc51101f39a59af195c3aa422b3e"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.460466 4979 generic.go:334] "Generic (PLEG): container finished" podID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" exitCode=0 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.460547 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerDied","Data":"64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.463020 4979 generic.go:334] "Generic (PLEG): container finished" podID="010e4389-4656-4460-9b4d-2a7a37660252" containerID="aa4d07bf279adaae12d8aad041308e01478fd7338e4e1c8dfe3e93d2a095aa65" exitCode=0 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.463066 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"010e4389-4656-4460-9b4d-2a7a37660252","Type":"ContainerDied","Data":"aa4d07bf279adaae12d8aad041308e01478fd7338e4e1c8dfe3e93d2a095aa65"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.466200 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanee53-account-delete-gpkgc" event={"ID":"a889a605-1dbe-4a0c-a91a-3a9d2153543a","Type":"ContainerStarted","Data":"0f6f3736f22758f16d22c55db1b531bcbd6407c3f1d52b34edd4bd38a33ba6be"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.486116 4979 generic.go:334] "Generic (PLEG): container finished" podID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" containerID="676d3e7b21b50404e78713930c4c90e6f133e77d13cd221bda1604c38311a704" exitCode=0 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.486191 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6","Type":"ContainerDied","Data":"676d3e7b21b50404e78713930c4c90e6f133e77d13cd221bda1604c38311a704"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.491363 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.494389 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement375f-account-delete-28b5l" event={"ID":"ec43b717-4a9b-47b0-9007-9589c7895729","Type":"ContainerStarted","Data":"1915aed98c7da8b97c132fd76f0542d8ba08dd838b88dd087f13a7386da9577f"} Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.496487 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "ef961b3d-b927-42e5-91b9-eb966bdee861" (UID: "ef961b3d-b927-42e5-91b9-eb966bdee861"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.534412 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.553922 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ef961b3d-b927-42e5-91b9-eb966bdee861-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.574914 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.577049 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" cmd=["/usr/local/bin/container-scripts/status_check.sh"] Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.577102 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-northd-0" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="ovn-northd" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.591731 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-db585f97f-d5777"] Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.592210 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-db585f97f-d5777" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-server" containerID="cri-o://0c6ee4d9251618d455bcbaeca27562a8a00ff00726d03c52d5bbc4be6e44f194" gracePeriod=30 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.592043 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/swift-proxy-db585f97f-d5777" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-httpd" containerID="cri-o://17d10021029714eaf81e19c13f5d0dfddee111a3c2ce627763576fdf61297d0c" gracePeriod=30 Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.656409 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.656484 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data podName:0fad7195-d19a-48ce-ad3d-4c67b53d2974 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.656468077 +0000 UTC m=+1518.930763881 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data") pod "rabbitmq-cell1-server-0" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974") : configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.803058 4979 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Dec 04 12:08:10 crc kubenswrapper[4979]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-04T12:08:08Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 04 12:08:10 crc kubenswrapper[4979]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 04 12:08:10 crc kubenswrapper[4979]: > execCommand=["/usr/share/ovn/scripts/ovn-ctl","stop_controller"] containerName="ovn-controller" pod="openstack/ovn-controller-x2hnf" message=< Dec 04 12:08:10 crc kubenswrapper[4979]: Exiting ovn-controller (1) [FAILED] Dec 04 12:08:10 crc kubenswrapper[4979]: Killing ovn-controller (1) [ OK ] Dec 04 12:08:10 crc kubenswrapper[4979]: 2025-12-04T12:08:08Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 04 12:08:10 crc kubenswrapper[4979]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 04 12:08:10 crc kubenswrapper[4979]: > Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.803098 4979 kuberuntime_container.go:691] "PreStop hook failed" err=< Dec 04 12:08:10 crc kubenswrapper[4979]: command '/usr/share/ovn/scripts/ovn-ctl stop_controller' exited with 137: 2025-12-04T12:08:08Z|00001|fatal_signal|WARN|terminating with signal 14 (Alarm clock) Dec 04 12:08:10 crc kubenswrapper[4979]: /etc/init.d/functions: line 589: 407 Alarm clock "$@" Dec 04 12:08:10 crc kubenswrapper[4979]: > pod="openstack/ovn-controller-x2hnf" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" containerID="cri-o://5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.803134 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-x2hnf" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" containerID="cri-o://5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" gracePeriod=27 Dec 04 12:08:10 crc kubenswrapper[4979]: I1204 12:08:10.804826 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-x2hnf" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" probeResult="failure" output="" Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.962649 4979 secret.go:188] Couldn't get secret openstack/cinder-scheduler-config-data: secret "cinder-scheduler-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.979504 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.979472853 +0000 UTC m=+1519.253768657 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data-custom" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scheduler-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.963150 4979 secret.go:188] Couldn't get secret openstack/cinder-config-data: secret "cinder-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.979623 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.979595537 +0000 UTC m=+1519.253891341 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-config-data" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.963208 4979 secret.go:188] Couldn't get secret openstack/cinder-scripts: secret "cinder-scripts" not found Dec 04 12:08:10 crc kubenswrapper[4979]: E1204 12:08:10.979665 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts podName:010e4389-4656-4460-9b4d-2a7a37660252 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.979655708 +0000 UTC m=+1519.253951512 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts") pod "cinder-scheduler-0" (UID: "010e4389-4656-4460-9b4d-2a7a37660252") : secret "cinder-scripts" not found Dec 04 12:08:11 crc kubenswrapper[4979]: E1204 12:08:11.172098 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca is running failed: container process not found" containerID="fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:11 crc kubenswrapper[4979]: E1204 12:08:11.182842 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca is running failed: container process not found" containerID="fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:11 crc kubenswrapper[4979]: E1204 12:08:11.195982 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca is running failed: container process not found" containerID="fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:11 crc kubenswrapper[4979]: E1204 12:08:11.196675 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca is running failed: container process not found" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" containerName="nova-cell0-conductor-conductor" Dec 04 12:08:11 crc kubenswrapper[4979]: E1204 12:08:11.261529 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod122bb56e_b8fc_4822_b6f5_6c91122ff4ee.slice/crio-fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb06a4c5f_10af_4f68_a303_f9342b97172b.slice/crio-conmon-5c5f1bce73c540715d7e8d457680551f4966428323cb84c1dbe1d33aa8ee2de6.scope\": RecentStats: unable to find data in memory cache]" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.480970 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.169:9292/healthcheck\": dial tcp 10.217.0.169:9292: connect: connection refused" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.481895 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-external-api-0" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.169:9292/healthcheck\": dial tcp 10.217.0.169:9292: connect: connection refused" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.530045 4979 generic.go:334] "Generic (PLEG): container finished" podID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerID="fe043f249efb9dfd8bb0e3a20f1360262cf5b2c026e1870681a1fabdae2c45d6" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.530141 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3791d23-03e1-4c8d-b44e-738eff9222d0","Type":"ContainerDied","Data":"fe043f249efb9dfd8bb0e3a20f1360262cf5b2c026e1870681a1fabdae2c45d6"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.537994 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-x2hnf_2853b847-4bf0-4d0d-af7b-ca8bc87a9c99/ovn-controller/0.log" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.538094 4979 generic.go:334] "Generic (PLEG): container finished" podID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerID="5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" exitCode=143 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.538147 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf" event={"ID":"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99","Type":"ContainerDied","Data":"5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.538173 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-x2hnf" event={"ID":"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99","Type":"ContainerDied","Data":"a0d3997c62f68fc25228cb0f599a184a35be015cfd2038e9541835c1f07bb64f"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.538185 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0d3997c62f68fc25228cb0f599a184a35be015cfd2038e9541835c1f07bb64f" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.545099 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6","Type":"ContainerDied","Data":"10aceaaa15915d853e102c8f274d765fecb48b39e94ae3626fc349a5d04aa448"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.545135 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10aceaaa15915d853e102c8f274d765fecb48b39e94ae3626fc349a5d04aa448" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.552522 4979 generic.go:334] "Generic (PLEG): container finished" podID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerID="063e5850e136541b04a3fbcd523818e0db023f1015cc97b6258d7d2440f0fc84" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.552619 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c","Type":"ContainerDied","Data":"063e5850e136541b04a3fbcd523818e0db023f1015cc97b6258d7d2440f0fc84"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.554324 4979 generic.go:334] "Generic (PLEG): container finished" podID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" containerID="fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.554389 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"122bb56e-b8fc-4822-b6f5-6c91122ff4ee","Type":"ContainerDied","Data":"fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.554405 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"122bb56e-b8fc-4822-b6f5-6c91122ff4ee","Type":"ContainerDied","Data":"af9f0679b273274bb1a02d208857e1b9f9cdec84ed8e963065c372d42f9bc905"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.554416 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af9f0679b273274bb1a02d208857e1b9f9cdec84ed8e963065c372d42f9bc905" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.556547 4979 generic.go:334] "Generic (PLEG): container finished" podID="010e4389-4656-4460-9b4d-2a7a37660252" containerID="66265992f836c6a9f8e585ba56a557ed7f384463e217621e5223701c105f7441" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.556627 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"010e4389-4656-4460-9b4d-2a7a37660252","Type":"ContainerDied","Data":"66265992f836c6a9f8e585ba56a557ed7f384463e217621e5223701c105f7441"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.556655 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"010e4389-4656-4460-9b4d-2a7a37660252","Type":"ContainerDied","Data":"a7d0f5cde648670e4314265bd9d2c17a2a777fd27c1c26186aae34f8afd33964"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.556665 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7d0f5cde648670e4314265bd9d2c17a2a777fd27c1c26186aae34f8afd33964" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.558397 4979 generic.go:334] "Generic (PLEG): container finished" podID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerID="dfee452098fffdbca65d5be95466eeb6448db8377ba02e117a1cd9cea7962b7f" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.558451 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5667db5f8d-q4ffn" event={"ID":"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd","Type":"ContainerDied","Data":"dfee452098fffdbca65d5be95466eeb6448db8377ba02e117a1cd9cea7962b7f"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.565397 4979 generic.go:334] "Generic (PLEG): container finished" podID="2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" containerID="dfb99c9b2fd24e29118b00f0d05832ad1a6f5468ea6b2c9b2e35da8ba48334d2" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.565446 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron47fe-account-delete-kqc4q" event={"ID":"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941","Type":"ContainerDied","Data":"dfb99c9b2fd24e29118b00f0d05832ad1a6f5468ea6b2c9b2e35da8ba48334d2"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.584014 4979 generic.go:334] "Generic (PLEG): container finished" podID="b06a4c5f-10af-4f68-a303-f9342b97172b" containerID="5c5f1bce73c540715d7e8d457680551f4966428323cb84c1dbe1d33aa8ee2de6" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.584074 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance2ebe-account-delete-wzsfb" event={"ID":"b06a4c5f-10af-4f68-a303-f9342b97172b","Type":"ContainerDied","Data":"5c5f1bce73c540715d7e8d457680551f4966428323cb84c1dbe1d33aa8ee2de6"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.595437 4979 generic.go:334] "Generic (PLEG): container finished" podID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerID="0c6ee4d9251618d455bcbaeca27562a8a00ff00726d03c52d5bbc4be6e44f194" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.595465 4979 generic.go:334] "Generic (PLEG): container finished" podID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerID="17d10021029714eaf81e19c13f5d0dfddee111a3c2ce627763576fdf61297d0c" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.595521 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-db585f97f-d5777" event={"ID":"ee632468-ca6f-48b2-abab-9cd939aafc00","Type":"ContainerDied","Data":"0c6ee4d9251618d455bcbaeca27562a8a00ff00726d03c52d5bbc4be6e44f194"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.595547 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-db585f97f-d5777" event={"ID":"ee632468-ca6f-48b2-abab-9cd939aafc00","Type":"ContainerDied","Data":"17d10021029714eaf81e19c13f5d0dfddee111a3c2ce627763576fdf61297d0c"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.597340 4979 generic.go:334] "Generic (PLEG): container finished" podID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerID="02024a485457048e2f54ce565223a36a682c9da052f1a487dff4004a22a382ee" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.597378 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d24f5a-ae53-44e3-bc4c-c56813f0f141","Type":"ContainerDied","Data":"02024a485457048e2f54ce565223a36a682c9da052f1a487dff4004a22a382ee"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.599313 4979 generic.go:334] "Generic (PLEG): container finished" podID="ec43b717-4a9b-47b0-9007-9589c7895729" containerID="e0d2fdaa377a63293816ad71e539f322b7ffe64f4aa7fab9b51cdf23832d0f21" exitCode=0 Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.599361 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement375f-account-delete-28b5l" event={"ID":"ec43b717-4a9b-47b0-9007-9589c7895729","Type":"ContainerDied","Data":"e0d2fdaa377a63293816ad71e539f322b7ffe64f4aa7fab9b51cdf23832d0f21"} Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.669104 4979 scope.go:117] "RemoveContainer" containerID="43237a8cc1d10bdde09a5a677588338264c1215b57330e0176e9bd4306ed5bf6" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.709476 4979 scope.go:117] "RemoveContainer" containerID="c86c01df6bfc54f8d0393df25c87596f615d07882e106e7f3454bf253a1dd6ef" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.796643 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57cfcdd946-gb75v" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:32808->10.217.0.159:9311: read: connection reset by peer" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.797370 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-57cfcdd946-gb75v" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.159:9311/healthcheck\": read tcp 10.217.0.2:32816->10.217.0.159:9311: read: connection reset by peer" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.799321 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.891463 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-nova-novncproxy-tls-certs\") pod \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.891589 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-vencrypt-tls-certs\") pod \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.891899 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-config-data\") pod \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.891960 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mztzj\" (UniqueName: \"kubernetes.io/projected/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-kube-api-access-mztzj\") pod \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.892000 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-combined-ca-bundle\") pod \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\" (UID: \"ac61c2b8-6dc7-41c5-a2e2-63759d5671b6\") " Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.911592 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-kube-api-access-mztzj" (OuterVolumeSpecName: "kube-api-access-mztzj") pod "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" (UID: "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6"). InnerVolumeSpecName "kube-api-access-mztzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.924563 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" (UID: "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.950004 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-config-data" (OuterVolumeSpecName: "config-data") pod "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" (UID: "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.960867 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-nova-novncproxy-tls-certs" (OuterVolumeSpecName: "nova-novncproxy-tls-certs") pod "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" (UID: "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6"). InnerVolumeSpecName "nova-novncproxy-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.965422 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-vencrypt-tls-certs" (OuterVolumeSpecName: "vencrypt-tls-certs") pod "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" (UID: "ac61c2b8-6dc7-41c5-a2e2-63759d5671b6"). InnerVolumeSpecName "vencrypt-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.994061 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.994105 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mztzj\" (UniqueName: \"kubernetes.io/projected/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-kube-api-access-mztzj\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.994119 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.994132 4979 reconciler_common.go:293] "Volume detached for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-nova-novncproxy-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:11 crc kubenswrapper[4979]: I1204 12:08:11.994144 4979 reconciler_common.go:293] "Volume detached for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6-vencrypt-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.117493 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.118349 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-central-agent" containerID="cri-o://2d030700ed1d0dceefff7c85f3b95944594dfdf24ef4f4fa9ee05c4e4c775576" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.118554 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="proxy-httpd" containerID="cri-o://c1c754fe678fbfa03811705e43d6a8012c1cb6310eb4b53a2378db11a7f2248c" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.118599 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="sg-core" containerID="cri-o://488ba15e74f2c14428d32f7d8b8d5f64b28117aabdfcb4b764a8f0e0a1923d0f" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.118636 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-notification-agent" containerID="cri-o://c4914f15e8bbbd431d76efb3ff6d2212902c6e00ddaf95b123156178bfa53c1c" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.141050 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.141358 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="13394b46-eca0-4f42-b3e7-88e20f0bb590" containerName="kube-state-metrics" containerID="cri-o://3300143eb0aae60225a16a5e6c14ef4fb760d477daea629b42dab44e38560ddc" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.474463 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.474703 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/memcached-0" podUID="f763e763-10df-408d-9452-84634d47b455" containerName="memcached" containerID="cri-o://b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.565354 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-hfg44"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.591353 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-hfg44"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.624928 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-4q48m"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.635376 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-4q48m"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.649405 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7c66cf746c-gkxxt"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.649699 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/keystone-7c66cf746c-gkxxt" podUID="983663bf-7439-4648-8237-1174fe609970" containerName="keystone-api" containerID="cri-o://91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1" gracePeriod=30 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.663735 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.664790 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.664960 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320 is running failed: container process not found" containerID="5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.668666 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.668731 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320 is running failed: container process not found" containerID="5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.678281 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.678355 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.678414 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320 is running failed: container process not found" containerID="5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" cmd=["/usr/local/bin/container-scripts/ovn_controller_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.678432 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320 is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-x2hnf" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.726360 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-zxkmj"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.750992 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.782754 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.783179 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-zxkmj"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.783215 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"e2d24f5a-ae53-44e3-bc4c-c56813f0f141","Type":"ContainerDied","Data":"a2ff341b559a4aa3d9cc2c4f5310f92c3beb84d8e92027d88499f5318b3249a4"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.783260 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2ff341b559a4aa3d9cc2c4f5310f92c3beb84d8e92027d88499f5318b3249a4" Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.823494 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:12 crc kubenswrapper[4979]: E1204 12:08:12.823919 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.837631 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ac67-account-create-update-ztshn"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.849866 4979 generic.go:334] "Generic (PLEG): container finished" podID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerID="c1697a681e440b41b665f6d4ba7c31e3a29aecbb77162bfc4b85770584ea8a4e" exitCode=0 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.849982 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cfcdd946-gb75v" event={"ID":"1ba50945-6ce9-4a71-9f58-e436b646c779","Type":"ContainerDied","Data":"c1697a681e440b41b665f6d4ba7c31e3a29aecbb77162bfc4b85770584ea8a4e"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.850417 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ac67-account-create-update-ztshn"] Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.857727 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c","Type":"ContainerDied","Data":"39910816ab33a8692857530e858fc019709d1d1558e496405b3d4da777c1dc83"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.857922 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39910816ab33a8692857530e858fc019709d1d1558e496405b3d4da777c1dc83" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.879631 4979 generic.go:334] "Generic (PLEG): container finished" podID="c472e576-897a-4ea6-912a-2d87b2f67077" containerID="0d566502255055e35e3f3b677f5eefb5b5ec440a9ad5528a85688965a96d78c3" exitCode=0 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.879727 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c472e576-897a-4ea6-912a-2d87b2f67077","Type":"ContainerDied","Data":"0d566502255055e35e3f3b677f5eefb5b5ec440a9ad5528a85688965a96d78c3"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.883910 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5667db5f8d-q4ffn" event={"ID":"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd","Type":"ContainerDied","Data":"43712bce63202df71bb07043fdbae8edb9815e8cab5ccc23b36112d41f7aa833"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.883948 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43712bce63202df71bb07043fdbae8edb9815e8cab5ccc23b36112d41f7aa833" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.894343 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d3791d23-03e1-4c8d-b44e-738eff9222d0","Type":"ContainerDied","Data":"0e7aa55bc3b8f0c516ee6b7c961e4d6cdb3d45ae3309ae2fb4bb62fad283ab6b"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.894544 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e7aa55bc3b8f0c516ee6b7c961e4d6cdb3d45ae3309ae2fb4bb62fad283ab6b" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.901960 4979 generic.go:334] "Generic (PLEG): container finished" podID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerID="3d54c5d9cb36dae2f6bcd838a3e0fddaf4c2d1f19e347ee987290f673c8efd37" exitCode=0 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.902063 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4ad703ff-c521-40f1-9fa4-09fca7726079","Type":"ContainerDied","Data":"3d54c5d9cb36dae2f6bcd838a3e0fddaf4c2d1f19e347ee987290f673c8efd37"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.902091 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"4ad703ff-c521-40f1-9fa4-09fca7726079","Type":"ContainerDied","Data":"bafb6eed795747c7239e30c7c1fd37c82ec310d9273eca36e1013feae0910bb7"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.902102 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bafb6eed795747c7239e30c7c1fd37c82ec310d9273eca36e1013feae0910bb7" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.912670 4979 generic.go:334] "Generic (PLEG): container finished" podID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerID="c1c754fe678fbfa03811705e43d6a8012c1cb6310eb4b53a2378db11a7f2248c" exitCode=0 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.912746 4979 generic.go:334] "Generic (PLEG): container finished" podID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerID="488ba15e74f2c14428d32f7d8b8d5f64b28117aabdfcb4b764a8f0e0a1923d0f" exitCode=2 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.912799 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerDied","Data":"c1c754fe678fbfa03811705e43d6a8012c1cb6310eb4b53a2378db11a7f2248c"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.912829 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerDied","Data":"488ba15e74f2c14428d32f7d8b8d5f64b28117aabdfcb4b764a8f0e0a1923d0f"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.915495 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-db585f97f-d5777" event={"ID":"ee632468-ca6f-48b2-abab-9cd939aafc00","Type":"ContainerDied","Data":"9a87d1ecab55532e46e04e3cfc19b09df18880b41260ad47483b6cbddc18770a"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.915528 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a87d1ecab55532e46e04e3cfc19b09df18880b41260ad47483b6cbddc18770a" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.916653 4979 generic.go:334] "Generic (PLEG): container finished" podID="13394b46-eca0-4f42-b3e7-88e20f0bb590" containerID="3300143eb0aae60225a16a5e6c14ef4fb760d477daea629b42dab44e38560ddc" exitCode=2 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.916734 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"13394b46-eca0-4f42-b3e7-88e20f0bb590","Type":"ContainerDied","Data":"3300143eb0aae60225a16a5e6c14ef4fb760d477daea629b42dab44e38560ddc"} Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.918349 4979 generic.go:334] "Generic (PLEG): container finished" podID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerID="640bdb1e4d1f6981f6d0fba2f1150c83d60fe36e91add11c47a642a8218b2507" exitCode=0 Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.918426 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 12:08:12 crc kubenswrapper[4979]: I1204 12:08:12.919075 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3238d4e5-e968-4d66-87d6-7397f48e9714","Type":"ContainerDied","Data":"640bdb1e4d1f6981f6d0fba2f1150c83d60fe36e91add11c47a642a8218b2507"} Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.107035 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="7130263f-15db-4a87-a078-9c346739eced" containerName="galera" containerID="cri-o://1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923" gracePeriod=30 Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.148686 4979 scope.go:117] "RemoveContainer" containerID="da2072151fb1ba04ce685c63911129fe34aff9fc376cba5805e8cd1d4a2e02b6" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.164071 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.189254 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-x2hnf_2853b847-4bf0-4d0d-af7b-ca8bc87a9c99/ovn-controller/0.log" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.189369 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.195110 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.201193 4979 scope.go:117] "RemoveContainer" containerID="0c220481bc2b2b19bd3e07a0f11e49e59396e88391fe837510160557d92051ac" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.209450 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.332941 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-ovn-controller-tls-certs\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.332997 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-log-ovn\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.333035 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnlch\" (UniqueName: \"kubernetes.io/projected/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-kube-api-access-hnlch\") pod \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.333060 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-combined-ca-bundle\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.333945 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.335315 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdps4\" (UniqueName: \"kubernetes.io/projected/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-kube-api-access-vdps4\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.335441 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-scripts\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.335490 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-combined-ca-bundle\") pod \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.335536 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-config-data\") pod \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\" (UID: \"122bb56e-b8fc-4822-b6f5-6c91122ff4ee\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.335586 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run-ovn\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.335643 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run\") pod \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\" (UID: \"2853b847-4bf0-4d0d-af7b-ca8bc87a9c99\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.336072 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run" (OuterVolumeSpecName: "var-run") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.336339 4979 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.336351 4979 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.336656 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-scripts" (OuterVolumeSpecName: "scripts") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.337697 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.357633 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-kube-api-access-vdps4" (OuterVolumeSpecName: "kube-api-access-vdps4") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "kube-api-access-vdps4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.357848 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-kube-api-access-hnlch" (OuterVolumeSpecName: "kube-api-access-hnlch") pod "122bb56e-b8fc-4822-b6f5-6c91122ff4ee" (UID: "122bb56e-b8fc-4822-b6f5-6c91122ff4ee"). InnerVolumeSpecName "kube-api-access-hnlch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.446813 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdps4\" (UniqueName: \"kubernetes.io/projected/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-kube-api-access-vdps4\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.446853 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.446866 4979 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.446878 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnlch\" (UniqueName: \"kubernetes.io/projected/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-kube-api-access-hnlch\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.532160 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.548925 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.583666 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-config-data" (OuterVolumeSpecName: "config-data") pod "122bb56e-b8fc-4822-b6f5-6c91122ff4ee" (UID: "122bb56e-b8fc-4822-b6f5-6c91122ff4ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.653418 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.657573 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "122bb56e-b8fc-4822-b6f5-6c91122ff4ee" (UID: "122bb56e-b8fc-4822-b6f5-6c91122ff4ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.700959 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.704347 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.713955 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.714379 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.714588 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.717587 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-ovn-controller-tls-certs" (OuterVolumeSpecName: "ovn-controller-tls-certs") pod "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" (UID: "2853b847-4bf0-4d0d-af7b-ca8bc87a9c99"). InnerVolumeSpecName "ovn-controller-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.725122 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.755517 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/122bb56e-b8fc-4822-b6f5-6c91122ff4ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.755555 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99-ovn-controller-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.760605 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.768082 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.783417 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.793459 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.797801 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.799419 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.807977 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.820153 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.820959 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.842369 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.847490 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857734 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-combined-ca-bundle\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857782 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwx8t\" (UniqueName: \"kubernetes.io/projected/d3791d23-03e1-4c8d-b44e-738eff9222d0-kube-api-access-kwx8t\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857810 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-config-data\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857835 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-combined-ca-bundle\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857856 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-public-tls-certs\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857876 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-scripts\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857916 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-log-httpd\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857939 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-default\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857970 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-internal-tls-certs\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.857987 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-scripts\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858004 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-scripts\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858030 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts\") pod \"010e4389-4656-4460-9b4d-2a7a37660252\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858047 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-config-data\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858066 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data\") pod \"010e4389-4656-4460-9b4d-2a7a37660252\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858084 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-httpd-run\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858104 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-logs\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858122 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgpkl\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-kube-api-access-lgpkl\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858139 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-internal-tls-certs\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858161 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-public-tls-certs\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858182 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zlfl\" (UniqueName: \"kubernetes.io/projected/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-kube-api-access-2zlfl\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858199 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-logs\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858215 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-galera-tls-certs\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858230 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-internal-tls-certs\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858253 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom\") pod \"010e4389-4656-4460-9b4d-2a7a37660252\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858283 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-generated\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858321 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-kolla-config\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858339 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-etc-swift\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858357 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-config-data\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858378 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-combined-ca-bundle\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858394 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-combined-ca-bundle\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858473 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgxkb\" (UniqueName: \"kubernetes.io/projected/010e4389-4656-4460-9b4d-2a7a37660252-kube-api-access-bgxkb\") pod \"010e4389-4656-4460-9b4d-2a7a37660252\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858495 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-config-data\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858511 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccfcl\" (UniqueName: \"kubernetes.io/projected/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-kube-api-access-ccfcl\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858529 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-httpd-run\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858553 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6zcf\" (UniqueName: \"kubernetes.io/projected/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-kube-api-access-w6zcf\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858577 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-public-tls-certs\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858605 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-run-httpd\") pod \"ee632468-ca6f-48b2-abab-9cd939aafc00\" (UID: \"ee632468-ca6f-48b2-abab-9cd939aafc00\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858640 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-logs\") pod \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\" (UID: \"000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858664 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-combined-ca-bundle\") pod \"010e4389-4656-4460-9b4d-2a7a37660252\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858699 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-operator-scripts\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858730 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/010e4389-4656-4460-9b4d-2a7a37660252-etc-machine-id\") pod \"010e4389-4656-4460-9b4d-2a7a37660252\" (UID: \"010e4389-4656-4460-9b4d-2a7a37660252\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858753 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"d3791d23-03e1-4c8d-b44e-738eff9222d0\" (UID: \"d3791d23-03e1-4c8d-b44e-738eff9222d0\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858766 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\" (UID: \"3cd5e649-996b-4558-b75d-b4cb3dfe2d4c\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858787 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.858804 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-combined-ca-bundle\") pod \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\" (UID: \"e2d24f5a-ae53-44e3-bc4c-c56813f0f141\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.860311 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.861864 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.870566 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-logs" (OuterVolumeSpecName: "logs") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.870728 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.877946 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.878090 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-kube-api-access-2zlfl" (OuterVolumeSpecName: "kube-api-access-2zlfl") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "kube-api-access-2zlfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.878172 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-logs" (OuterVolumeSpecName: "logs") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.878846 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.879409 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/010e4389-4656-4460-9b4d-2a7a37660252-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "010e4389-4656-4460-9b4d-2a7a37660252" (UID: "010e4389-4656-4460-9b4d-2a7a37660252"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.894979 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-logs" (OuterVolumeSpecName: "logs") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.899191 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.899658 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.900601 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.904552 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.905006 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3791d23-03e1-4c8d-b44e-738eff9222d0-kube-api-access-kwx8t" (OuterVolumeSpecName: "kube-api-access-kwx8t") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "kube-api-access-kwx8t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.917100 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.918798 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts" (OuterVolumeSpecName: "scripts") pod "010e4389-4656-4460-9b4d-2a7a37660252" (UID: "010e4389-4656-4460-9b4d-2a7a37660252"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.919385 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-scripts" (OuterVolumeSpecName: "scripts") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.921234 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-kube-api-access-w6zcf" (OuterVolumeSpecName: "kube-api-access-w6zcf") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "kube-api-access-w6zcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.921270 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/010e4389-4656-4460-9b4d-2a7a37660252-kube-api-access-bgxkb" (OuterVolumeSpecName: "kube-api-access-bgxkb") pod "010e4389-4656-4460-9b4d-2a7a37660252" (UID: "010e4389-4656-4460-9b4d-2a7a37660252"). InnerVolumeSpecName "kube-api-access-bgxkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.932594 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-scripts" (OuterVolumeSpecName: "scripts") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.949294 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.950446 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.952920 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-kube-api-access-ccfcl" (OuterVolumeSpecName: "kube-api-access-ccfcl") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "kube-api-access-ccfcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.953006 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-kube-api-access-lgpkl" (OuterVolumeSpecName: "kube-api-access-lgpkl") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "kube-api-access-lgpkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.956384 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder37f5-account-delete-4wn9x" event={"ID":"1341b825-bf58-4a12-af3e-40411a861e8f","Type":"ContainerStarted","Data":"02302b47b9f3434931fdab45d8a9de412215496601beedc94b30eaf534866d88"} Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.957344 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder37f5-account-delete-4wn9x" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.957555 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960065 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06a4c5f-10af-4f68-a303-f9342b97172b-operator-scripts\") pod \"b06a4c5f-10af-4f68-a303-f9342b97172b\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960163 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-combined-ca-bundle\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960382 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d6dk\" (UniqueName: \"kubernetes.io/projected/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-api-access-6d6dk\") pod \"13394b46-eca0-4f42-b3e7-88e20f0bb590\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960412 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-public-tls-certs\") pod \"c472e576-897a-4ea6-912a-2d87b2f67077\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960449 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960466 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csgn7\" (UniqueName: \"kubernetes.io/projected/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-kube-api-access-csgn7\") pod \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960546 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-scripts\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960576 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-operator-scripts\") pod \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\" (UID: \"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960603 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9k4\" (UniqueName: \"kubernetes.io/projected/b06a4c5f-10af-4f68-a303-f9342b97172b-kube-api-access-2w9k4\") pod \"b06a4c5f-10af-4f68-a303-f9342b97172b\" (UID: \"b06a4c5f-10af-4f68-a303-f9342b97172b\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960678 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data-custom\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960708 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-combined-ca-bundle\") pod \"13394b46-eca0-4f42-b3e7-88e20f0bb590\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960736 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-combined-ca-bundle\") pod \"3238d4e5-e968-4d66-87d6-7397f48e9714\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960754 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65zkd\" (UniqueName: \"kubernetes.io/projected/c472e576-897a-4ea6-912a-2d87b2f67077-kube-api-access-65zkd\") pod \"c472e576-897a-4ea6-912a-2d87b2f67077\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960771 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-nova-metadata-tls-certs\") pod \"3238d4e5-e968-4d66-87d6-7397f48e9714\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960796 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c472e576-897a-4ea6-912a-2d87b2f67077-logs\") pod \"c472e576-897a-4ea6-912a-2d87b2f67077\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960874 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-combined-ca-bundle\") pod \"c472e576-897a-4ea6-912a-2d87b2f67077\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960919 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-combined-ca-bundle\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960942 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ba50945-6ce9-4a71-9f58-e436b646c779-logs\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.960958 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqtd4\" (UniqueName: \"kubernetes.io/projected/1ba50945-6ce9-4a71-9f58-e436b646c779-kube-api-access-gqtd4\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961000 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-config-data\") pod \"3238d4e5-e968-4d66-87d6-7397f48e9714\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961020 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961035 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ad703ff-c521-40f1-9fa4-09fca7726079-etc-machine-id\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961058 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-config\") pod \"13394b46-eca0-4f42-b3e7-88e20f0bb590\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961084 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-internal-tls-certs\") pod \"c472e576-897a-4ea6-912a-2d87b2f67077\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961114 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-public-tls-certs\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961565 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3238d4e5-e968-4d66-87d6-7397f48e9714-logs\") pod \"3238d4e5-e968-4d66-87d6-7397f48e9714\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961615 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-659ct\" (UniqueName: \"kubernetes.io/projected/3238d4e5-e968-4d66-87d6-7397f48e9714-kube-api-access-659ct\") pod \"3238d4e5-e968-4d66-87d6-7397f48e9714\" (UID: \"3238d4e5-e968-4d66-87d6-7397f48e9714\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961641 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data-custom\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961711 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-internal-tls-certs\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.961982 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-public-tls-certs\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.962133 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-certs\") pod \"13394b46-eca0-4f42-b3e7-88e20f0bb590\" (UID: \"13394b46-eca0-4f42-b3e7-88e20f0bb590\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.962179 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxptx\" (UniqueName: \"kubernetes.io/projected/4ad703ff-c521-40f1-9fa4-09fca7726079-kube-api-access-cxptx\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.962207 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-internal-tls-certs\") pod \"1ba50945-6ce9-4a71-9f58-e436b646c779\" (UID: \"1ba50945-6ce9-4a71-9f58-e436b646c779\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.962313 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ad703ff-c521-40f1-9fa4-09fca7726079-logs\") pod \"4ad703ff-c521-40f1-9fa4-09fca7726079\" (UID: \"4ad703ff-c521-40f1-9fa4-09fca7726079\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.962402 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-config-data\") pod \"c472e576-897a-4ea6-912a-2d87b2f67077\" (UID: \"c472e576-897a-4ea6-912a-2d87b2f67077\") " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.962987 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963003 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963017 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963025 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963035 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963044 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgpkl\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-kube-api-access-lgpkl\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963054 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zlfl\" (UniqueName: \"kubernetes.io/projected/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-kube-api-access-2zlfl\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963062 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963092 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963110 4979 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963125 4979 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/ee632468-ca6f-48b2-abab-9cd939aafc00-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963136 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccfcl\" (UniqueName: \"kubernetes.io/projected/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-kube-api-access-ccfcl\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963145 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgxkb\" (UniqueName: \"kubernetes.io/projected/010e4389-4656-4460-9b4d-2a7a37660252-kube-api-access-bgxkb\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963154 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963162 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6zcf\" (UniqueName: \"kubernetes.io/projected/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-kube-api-access-w6zcf\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963170 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963178 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963187 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963196 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/010e4389-4656-4460-9b4d-2a7a37660252-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963247 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963261 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963270 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwx8t\" (UniqueName: \"kubernetes.io/projected/d3791d23-03e1-4c8d-b44e-738eff9222d0-kube-api-access-kwx8t\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963281 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ee632468-ca6f-48b2-abab-9cd939aafc00-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.963320 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d3791d23-03e1-4c8d-b44e-738eff9222d0-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.964539 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4ad703ff-c521-40f1-9fa4-09fca7726079-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.967246 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c472e576-897a-4ea6-912a-2d87b2f67077-logs" (OuterVolumeSpecName: "logs") pod "c472e576-897a-4ea6-912a-2d87b2f67077" (UID: "c472e576-897a-4ea6-912a-2d87b2f67077"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.967486 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b06a4c5f-10af-4f68-a303-f9342b97172b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b06a4c5f-10af-4f68-a303-f9342b97172b" (UID: "b06a4c5f-10af-4f68-a303-f9342b97172b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.967901 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.975618 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ba50945-6ce9-4a71-9f58-e436b646c779-logs" (OuterVolumeSpecName: "logs") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.986210 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" (UID: "2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.992037 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-57cfcdd946-gb75v" event={"ID":"1ba50945-6ce9-4a71-9f58-e436b646c779","Type":"ContainerDied","Data":"7852e4534fa46cd05339048e17823fa84f59d75ecdf087a64e4cae75ed60dd30"} Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.992105 4979 scope.go:117] "RemoveContainer" containerID="c1697a681e440b41b665f6d4ba7c31e3a29aecbb77162bfc4b85770584ea8a4e" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.992351 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-57cfcdd946-gb75v" Dec 04 12:08:13 crc kubenswrapper[4979]: I1204 12:08:13.993762 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3238d4e5-e968-4d66-87d6-7397f48e9714-logs" (OuterVolumeSpecName: "logs") pod "3238d4e5-e968-4d66-87d6-7397f48e9714" (UID: "3238d4e5-e968-4d66-87d6-7397f48e9714"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:13.996595 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ad703ff-c521-40f1-9fa4-09fca7726079-logs" (OuterVolumeSpecName: "logs") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:13.997642 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance2ebe-account-delete-wzsfb" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:13.997633 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance2ebe-account-delete-wzsfb" event={"ID":"b06a4c5f-10af-4f68-a303-f9342b97172b","Type":"ContainerDied","Data":"a7f3cfcc6ea9f0d1247475acda7e385c450f0f81f09b257c0d8728723a8cc66c"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:13.997756 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7f3cfcc6ea9f0d1247475acda7e385c450f0f81f09b257c0d8728723a8cc66c" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.003215 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.003228 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"3238d4e5-e968-4d66-87d6-7397f48e9714","Type":"ContainerDied","Data":"011026b7a80047db82851700f8e2c0e45e41fd86dc23c3f536e6c524c19a4670"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.005754 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron47fe-account-delete-kqc4q" event={"ID":"2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941","Type":"ContainerDied","Data":"3a7a77cfb23ce2082033ee114f5048e53209c125345774e7c7842daafb4fddd4"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.005781 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a7a77cfb23ce2082033ee114f5048e53209c125345774e7c7842daafb4fddd4" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.005831 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron47fe-account-delete-kqc4q" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.006792 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ba50945-6ce9-4a71-9f58-e436b646c779-kube-api-access-gqtd4" (OuterVolumeSpecName: "kube-api-access-gqtd4") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "kube-api-access-gqtd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.010291 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ad703ff-c521-40f1-9fa4-09fca7726079-kube-api-access-cxptx" (OuterVolumeSpecName: "kube-api-access-cxptx") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "kube-api-access-cxptx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.010479 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "010e4389-4656-4460-9b4d-2a7a37660252" (UID: "010e4389-4656-4460-9b4d-2a7a37660252"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.010649 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c472e576-897a-4ea6-912a-2d87b2f67077-kube-api-access-65zkd" (OuterVolumeSpecName: "kube-api-access-65zkd") pod "c472e576-897a-4ea6-912a-2d87b2f67077" (UID: "c472e576-897a-4ea6-912a-2d87b2f67077"). InnerVolumeSpecName "kube-api-access-65zkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.020118 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanee53-account-delete-gpkgc" event={"ID":"a889a605-1dbe-4a0c-a91a-3a9d2153543a","Type":"ContainerStarted","Data":"3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.020707 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicanee53-account-delete-gpkgc" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.023937 4979 generic.go:334] "Generic (PLEG): container finished" podID="f763e763-10df-408d-9452-84634d47b455" containerID="b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c" exitCode=0 Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.024034 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f763e763-10df-408d-9452-84634d47b455","Type":"ContainerDied","Data":"b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.024061 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"f763e763-10df-408d-9452-84634d47b455","Type":"ContainerDied","Data":"298d0f7088901bd1bf19e095b09802198329ed8cbbe172cad02e5d04b8488026"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.024382 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.026669 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.028928 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-scripts" (OuterVolumeSpecName: "scripts") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.030008 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.030031 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c472e576-897a-4ea6-912a-2d87b2f67077","Type":"ContainerDied","Data":"5b251dc57aafe3af2839d9c2c0ebaecf76adeef682678c03c10153636b5311df"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.032821 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02b03-account-delete-hf9n8" event={"ID":"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4","Type":"ContainerStarted","Data":"7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.033448 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell02b03-account-delete-hf9n8" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.034864 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement375f-account-delete-28b5l" event={"ID":"ec43b717-4a9b-47b0-9007-9589c7895729","Type":"ContainerDied","Data":"1915aed98c7da8b97c132fd76f0542d8ba08dd838b88dd087f13a7386da9577f"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.034885 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1915aed98c7da8b97c132fd76f0542d8ba08dd838b88dd087f13a7386da9577f" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.034944 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement375f-account-delete-28b5l" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.039045 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic904-account-delete-cvwnr" event={"ID":"f9b386d0-b110-48a9-b6fe-377f0faeedc0","Type":"ContainerStarted","Data":"3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.039644 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapic904-account-delete-cvwnr" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.043153 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"13394b46-eca0-4f42-b3e7-88e20f0bb590","Type":"ContainerDied","Data":"93d75a52157d4e3ace8193c986b29a1deea9b116c20ff06d193387a34b3650ed"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.043220 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.049834 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "mysql-db") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.051554 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-api-access-6d6dk" (OuterVolumeSpecName: "kube-api-access-6d6dk") pod "13394b46-eca0-4f42-b3e7-88e20f0bb590" (UID: "13394b46-eca0-4f42-b3e7-88e20f0bb590"). InnerVolumeSpecName "kube-api-access-6d6dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054256 4979 generic.go:334] "Generic (PLEG): container finished" podID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerID="2d030700ed1d0dceefff7c85f3b95944594dfdf24ef4f4fa9ee05c4e4c775576" exitCode=0 Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054378 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5667db5f8d-q4ffn" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054389 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerDied","Data":"2d030700ed1d0dceefff7c85f3b95944594dfdf24ef4f4fa9ee05c4e4c775576"} Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054541 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-db585f97f-d5777" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054596 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054607 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054627 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054659 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054676 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054684 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.054860 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-x2hnf" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064098 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-memcached-tls-certs\") pod \"f763e763-10df-408d-9452-84634d47b455\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064149 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-config-data\") pod \"f763e763-10df-408d-9452-84634d47b455\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064231 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lb8d9\" (UniqueName: \"kubernetes.io/projected/ec43b717-4a9b-47b0-9007-9589c7895729-kube-api-access-lb8d9\") pod \"ec43b717-4a9b-47b0-9007-9589c7895729\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064376 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec43b717-4a9b-47b0-9007-9589c7895729-operator-scripts\") pod \"ec43b717-4a9b-47b0-9007-9589c7895729\" (UID: \"ec43b717-4a9b-47b0-9007-9589c7895729\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064446 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-combined-ca-bundle\") pod \"f763e763-10df-408d-9452-84634d47b455\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064672 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-kolla-config\") pod \"f763e763-10df-408d-9452-84634d47b455\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064664 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b06a4c5f-10af-4f68-a303-f9342b97172b-kube-api-access-2w9k4" (OuterVolumeSpecName: "kube-api-access-2w9k4") pod "b06a4c5f-10af-4f68-a303-f9342b97172b" (UID: "b06a4c5f-10af-4f68-a303-f9342b97172b"). InnerVolumeSpecName "kube-api-access-2w9k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.064868 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wszst\" (UniqueName: \"kubernetes.io/projected/f763e763-10df-408d-9452-84634d47b455-kube-api-access-wszst\") pod \"f763e763-10df-408d-9452-84634d47b455\" (UID: \"f763e763-10df-408d-9452-84634d47b455\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.066869 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-config-data" (OuterVolumeSpecName: "config-data") pod "f763e763-10df-408d-9452-84634d47b455" (UID: "f763e763-10df-408d-9452-84634d47b455"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.068284 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec43b717-4a9b-47b0-9007-9589c7895729-kube-api-access-lb8d9" (OuterVolumeSpecName: "kube-api-access-lb8d9") pod "ec43b717-4a9b-47b0-9007-9589c7895729" (UID: "ec43b717-4a9b-47b0-9007-9589c7895729"). InnerVolumeSpecName "kube-api-access-lb8d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.068586 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec43b717-4a9b-47b0-9007-9589c7895729-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec43b717-4a9b-47b0-9007-9589c7895729" (UID: "ec43b717-4a9b-47b0-9007-9589c7895729"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.068908 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "f763e763-10df-408d-9452-84634d47b455" (UID: "f763e763-10df-408d-9452-84634d47b455"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069151 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4ad703ff-c521-40f1-9fa4-09fca7726079-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069312 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b06a4c5f-10af-4f68-a303-f9342b97172b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069327 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d6dk\" (UniqueName: \"kubernetes.io/projected/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-api-access-6d6dk\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069352 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069365 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069380 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9k4\" (UniqueName: \"kubernetes.io/projected/b06a4c5f-10af-4f68-a303-f9342b97172b-kube-api-access-2w9k4\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069391 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069403 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65zkd\" (UniqueName: \"kubernetes.io/projected/c472e576-897a-4ea6-912a-2d87b2f67077-kube-api-access-65zkd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069417 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c472e576-897a-4ea6-912a-2d87b2f67077-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069430 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lb8d9\" (UniqueName: \"kubernetes.io/projected/ec43b717-4a9b-47b0-9007-9589c7895729-kube-api-access-lb8d9\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069443 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqtd4\" (UniqueName: \"kubernetes.io/projected/1ba50945-6ce9-4a71-9f58-e436b646c779-kube-api-access-gqtd4\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069455 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ba50945-6ce9-4a71-9f58-e436b646c779-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069466 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4ad703ff-c521-40f1-9fa4-09fca7726079-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069478 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec43b717-4a9b-47b0-9007-9589c7895729-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069505 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069518 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3238d4e5-e968-4d66-87d6-7397f48e9714-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069530 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069573 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069597 4979 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f763e763-10df-408d-9452-84634d47b455-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069611 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.069624 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxptx\" (UniqueName: \"kubernetes.io/projected/4ad703ff-c521-40f1-9fa4-09fca7726079-kube-api-access-cxptx\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.069677 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.069726 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.569708388 +0000 UTC m=+1518.844004282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.094192 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.094219 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3238d4e5-e968-4d66-87d6-7397f48e9714-kube-api-access-659ct" (OuterVolumeSpecName: "kube-api-access-659ct") pod "3238d4e5-e968-4d66-87d6-7397f48e9714" (UID: "3238d4e5-e968-4d66-87d6-7397f48e9714"). InnerVolumeSpecName "kube-api-access-659ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.094266 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-scripts" (OuterVolumeSpecName: "scripts") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.094270 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-kube-api-access-csgn7" (OuterVolumeSpecName: "kube-api-access-csgn7") pod "2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" (UID: "2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941"). InnerVolumeSpecName "kube-api-access-csgn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.112187 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novacell02b03-account-delete-hf9n8" podStartSLOduration=7.112168901 podStartE2EDuration="7.112168901s" podCreationTimestamp="2025-12-04 12:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:08:14.104534316 +0000 UTC m=+1518.378830130" watchObservedRunningTime="2025-12-04 12:08:14.112168901 +0000 UTC m=+1518.386464705" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.115451 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f763e763-10df-408d-9452-84634d47b455-kube-api-access-wszst" (OuterVolumeSpecName: "kube-api-access-wszst") pod "f763e763-10df-408d-9452-84634d47b455" (UID: "f763e763-10df-408d-9452-84634d47b455"). InnerVolumeSpecName "kube-api-access-wszst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.152178 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.178963 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/novaapic904-account-delete-cvwnr" podStartSLOduration=7.178944119 podStartE2EDuration="7.178944119s" podCreationTimestamp="2025-12-04 12:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:08:14.152446906 +0000 UTC m=+1518.426742700" watchObservedRunningTime="2025-12-04 12:08:14.178944119 +0000 UTC m=+1518.453239923" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.195244 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.195332 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.69530974 +0000 UTC m=+1518.969605544 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.197173 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.197202 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.197246 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.697233261 +0000 UTC m=+1518.971529065 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.197264 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:14.697257742 +0000 UTC m=+1518.971553546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.197554 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wszst\" (UniqueName: \"kubernetes.io/projected/f763e763-10df-408d-9452-84634d47b455-kube-api-access-wszst\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.197575 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csgn7\" (UniqueName: \"kubernetes.io/projected/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941-kube-api-access-csgn7\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.197585 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.197611 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.197622 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.197633 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-659ct\" (UniqueName: \"kubernetes.io/projected/3238d4e5-e968-4d66-87d6-7397f48e9714-kube-api-access-659ct\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.221089 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbicanee53-account-delete-gpkgc" podStartSLOduration=7.221071053 podStartE2EDuration="7.221071053s" podCreationTimestamp="2025-12-04 12:08:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:08:14.219902372 +0000 UTC m=+1518.494198196" watchObservedRunningTime="2025-12-04 12:08:14.221071053 +0000 UTC m=+1518.495366867" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.223375 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder37f5-account-delete-4wn9x" podStartSLOduration=8.223361695 podStartE2EDuration="8.223361695s" podCreationTimestamp="2025-12-04 12:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:08:14.189131333 +0000 UTC m=+1518.463427137" watchObservedRunningTime="2025-12-04 12:08:14.223361695 +0000 UTC m=+1518.497657499" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.236423 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.199:3000/\": dial tcp 10.217.0.199:3000: connect: connection refused" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.244222 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09068fde-06eb-4075-b4aa-78c8de781ced" path="/var/lib/kubelet/pods/09068fde-06eb-4075-b4aa-78c8de781ced/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.253969 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b" path="/var/lib/kubelet/pods/30d3416a-9f5d-40ab-a7d9-2ba6299a5f8b/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.254878 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35e82e89-f20a-451f-b992-5b07d77b1e14" path="/var/lib/kubelet/pods/35e82e89-f20a-451f-b992-5b07d77b1e14/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.256138 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e3201e7-fdd0-49c2-93bc-96211c54e55a" path="/var/lib/kubelet/pods/8e3201e7-fdd0-49c2-93bc-96211c54e55a/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.257081 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" path="/var/lib/kubelet/pods/ac61c2b8-6dc7-41c5-a2e2-63759d5671b6/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.258727 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" path="/var/lib/kubelet/pods/bdd8e97d-9bc0-4e10-984b-7d92698e8bd6/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.259807 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" path="/var/lib/kubelet/pods/ef961b3d-b927-42e5-91b9-eb966bdee861/volumes" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.300798 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-config-data: configmap "rabbitmq-config-data" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.300887 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data podName:0e592d2c-bb50-44b9-bd85-fdf94ee1ac48 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:22.300859701 +0000 UTC m=+1526.575155595 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data") pod "rabbitmq-server-0" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48") : configmap "rabbitmq-config-data" not found Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.324717 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.343557 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.354616 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.388374 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-config-data" (OuterVolumeSpecName: "config-data") pod "c472e576-897a-4ea6-912a-2d87b2f67077" (UID: "c472e576-897a-4ea6-912a-2d87b2f67077"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.435437 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.438098 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.438118 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.438130 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.438140 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.438150 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.446740 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.447715 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "13394b46-eca0-4f42-b3e7-88e20f0bb590" (UID: "13394b46-eca0-4f42-b3e7-88e20f0bb590"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.475465 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.490853 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.492134 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.493370 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.493491 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="77ca261e-57dc-4b57-a80b-e7def9b113d5" containerName="nova-scheduler-scheduler" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.514511 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-config-data" (OuterVolumeSpecName: "config-data") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.520731 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3238d4e5-e968-4d66-87d6-7397f48e9714" (UID: "3238d4e5-e968-4d66-87d6-7397f48e9714"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.522459 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "3238d4e5-e968-4d66-87d6-7397f48e9714" (UID: "3238d4e5-e968-4d66-87d6-7397f48e9714"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.524417 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.525456 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.526519 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.526601 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell1-conductor-0" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerName="nova-cell1-conductor-conductor" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.541707 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f763e763-10df-408d-9452-84634d47b455" (UID: "f763e763-10df-408d-9452-84634d47b455"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.546606 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.550016 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.550129 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.550191 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.550282 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.550377 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.550441 4979 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.597873 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.102:5671: connect: connection refused" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.604701 4979 scope.go:117] "RemoveContainer" containerID="f79822b9ecbd0ae0552d4c0bef356ac9d8b4b613b0aa7af9b7905e5627f9d542" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.616456 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.616713 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d5e30fc1-acc8-448a-a9e8-490bda22e24e/ovn-northd/0.log" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.616794 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.629019 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-certs" (OuterVolumeSpecName: "kube-state-metrics-tls-certs") pod "13394b46-eca0-4f42-b3e7-88e20f0bb590" (UID: "13394b46-eca0-4f42-b3e7-88e20f0bb590"). InnerVolumeSpecName "kube-state-metrics-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.630613 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.103:5671: connect: connection refused" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.631764 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.631818 4979 scope.go:117] "RemoveContainer" containerID="640bdb1e4d1f6981f6d0fba2f1150c83d60fe36e91add11c47a642a8218b2507" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.634335 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.647293 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-x2hnf"] Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.647452 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data" (OuterVolumeSpecName: "config-data") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.652532 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-config\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.652683 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-metrics-certs-tls-certs\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.652764 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-combined-ca-bundle\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.653478 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-rundir\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.653618 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-northd-tls-certs\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.654552 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrtz8\" (UniqueName: \"kubernetes.io/projected/d5e30fc1-acc8-448a-a9e8-490bda22e24e-kube-api-access-zrtz8\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.655427 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-scripts\") pod \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\" (UID: \"d5e30fc1-acc8-448a-a9e8-490bda22e24e\") " Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.656024 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.656104 4979 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.656162 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.656458 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.656525 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:08:15.656507166 +0000 UTC m=+1519.930802970 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.653804 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-x2hnf"] Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.656920 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.660593 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-rundir" (OuterVolumeSpecName: "ovn-rundir") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "ovn-rundir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.670277 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-config" (OuterVolumeSpecName: "config") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.672359 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-memcached-tls-certs" (OuterVolumeSpecName: "memcached-tls-certs") pod "f763e763-10df-408d-9452-84634d47b455" (UID: "f763e763-10df-408d-9452-84634d47b455"). InnerVolumeSpecName "memcached-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.673515 4979 scope.go:117] "RemoveContainer" containerID="601adae13b36c5015c0e550950004699f70cac5db6ce0564490dc4284b7bfc50" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.673756 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-config-data" (OuterVolumeSpecName: "config-data") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.685735 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-scripts" (OuterVolumeSpecName: "scripts") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.691323 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5e30fc1-acc8-448a-a9e8-490bda22e24e-kube-api-access-zrtz8" (OuterVolumeSpecName: "kube-api-access-zrtz8") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "kube-api-access-zrtz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.697253 4979 scope.go:117] "RemoveContainer" containerID="b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.698442 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c472e576-897a-4ea6-912a-2d87b2f67077" (UID: "c472e576-897a-4ea6-912a-2d87b2f67077"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.700504 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.709691 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-config-data" (OuterVolumeSpecName: "config-data") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.732636 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-config" (OuterVolumeSpecName: "kube-state-metrics-tls-config") pod "13394b46-eca0-4f42-b3e7-88e20f0bb590" (UID: "13394b46-eca0-4f42-b3e7-88e20f0bb590"). InnerVolumeSpecName "kube-state-metrics-tls-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.734181 4979 scope.go:117] "RemoveContainer" containerID="b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.734440 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c472e576-897a-4ea6-912a-2d87b2f67077" (UID: "c472e576-897a-4ea6-912a-2d87b2f67077"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.734832 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c\": container with ID starting with b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c not found: ID does not exist" containerID="b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.734869 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c"} err="failed to get container status \"b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c\": rpc error: code = NotFound desc = could not find container \"b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c\": container with ID starting with b24f469bf4d3b630bbb03c8a684f48b803676f1f99c436b36c16059ba7e9c03c not found: ID does not exist" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.734898 4979 scope.go:117] "RemoveContainer" containerID="0d566502255055e35e3f3b677f5eefb5b5ec440a9ad5528a85688965a96d78c3" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.737621 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-config-data" (OuterVolumeSpecName: "config-data") pod "3238d4e5-e968-4d66-87d6-7397f48e9714" (UID: "3238d4e5-e968-4d66-87d6-7397f48e9714"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.754437 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757858 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3238d4e5-e968-4d66-87d6-7397f48e9714-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757894 4979 reconciler_common.go:293] "Volume detached for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/13394b46-eca0-4f42-b3e7-88e20f0bb590-kube-state-metrics-tls-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757907 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757919 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-rundir\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757930 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757941 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757952 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757962 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrtz8\" (UniqueName: \"kubernetes.io/projected/d5e30fc1-acc8-448a-a9e8-490bda22e24e-kube-api-access-zrtz8\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757974 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757984 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.757994 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.758003 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.758014 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d5e30fc1-acc8-448a-a9e8-490bda22e24e-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.758027 4979 reconciler_common.go:293] "Volume detached for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/f763e763-10df-408d-9452-84634d47b455-memcached-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758116 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758166 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:08:15.758148793 +0000 UTC m=+1520.032444597 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758628 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758654 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:15.758646526 +0000 UTC m=+1520.032942330 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758688 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758705 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:15.758699817 +0000 UTC m=+1520.032995621 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758727 4979 configmap.go:193] Couldn't get configMap openstack/rabbitmq-cell1-config-data: configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:14 crc kubenswrapper[4979]: E1204 12:08:14.758743 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data podName:0fad7195-d19a-48ce-ad3d-4c67b53d2974 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:22.758737878 +0000 UTC m=+1527.033033682 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data") pod "rabbitmq-cell1-server-0" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974") : configmap "rabbitmq-cell1-config-data" not found Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.764445 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c472e576-897a-4ea6-912a-2d87b2f67077" (UID: "c472e576-897a-4ea6-912a-2d87b2f67077"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.764487 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.764625 4979 scope.go:117] "RemoveContainer" containerID="ea884876a5d807dfae591e01b960188f749527339aa8424bc250430bc801a04f" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.785354 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" (UID: "3cd5e649-996b-4558-b75d-b4cb3dfe2d4c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.785430 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.788356 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.802106 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ee632468-ca6f-48b2-abab-9cd939aafc00" (UID: "ee632468-ca6f-48b2-abab-9cd939aafc00"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.811551 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "4ad703ff-c521-40f1-9fa4-09fca7726079" (UID: "4ad703ff-c521-40f1-9fa4-09fca7726079"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.827014 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-config-data" (OuterVolumeSpecName: "config-data") pod "e2d24f5a-ae53-44e3-bc4c-c56813f0f141" (UID: "e2d24f5a-ae53-44e3-bc4c-c56813f0f141"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.834261 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "d3791d23-03e1-4c8d-b44e-738eff9222d0" (UID: "d3791d23-03e1-4c8d-b44e-738eff9222d0"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.838276 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "010e4389-4656-4460-9b4d-2a7a37660252" (UID: "010e4389-4656-4460-9b4d-2a7a37660252"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.853773 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860221 4979 scope.go:117] "RemoveContainer" containerID="3300143eb0aae60225a16a5e6c14ef4fb760d477daea629b42dab44e38560ddc" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860527 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860763 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860775 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860786 4979 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d3791d23-03e1-4c8d-b44e-738eff9222d0-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860831 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee632468-ca6f-48b2-abab-9cd939aafc00-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860842 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2d24f5a-ae53-44e3-bc4c-c56813f0f141-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860853 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860863 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860914 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c472e576-897a-4ea6-912a-2d87b2f67077-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860925 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ad703ff-c521-40f1-9fa4-09fca7726079-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.860939 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.868693 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data" (OuterVolumeSpecName: "config-data") pod "010e4389-4656-4460-9b4d-2a7a37660252" (UID: "010e4389-4656-4460-9b4d-2a7a37660252"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.870366 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data" (OuterVolumeSpecName: "config-data") pod "1ba50945-6ce9-4a71-9f58-e436b646c779" (UID: "1ba50945-6ce9-4a71-9f58-e436b646c779"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.891787 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.930499 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-northd-tls-certs" (OuterVolumeSpecName: "ovn-northd-tls-certs") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "ovn-northd-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.932527 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-metrics-certs-tls-certs" (OuterVolumeSpecName: "metrics-certs-tls-certs") pod "d5e30fc1-acc8-448a-a9e8-490bda22e24e" (UID: "d5e30fc1-acc8-448a-a9e8-490bda22e24e"). InnerVolumeSpecName "metrics-certs-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.966485 4979 reconciler_common.go:293] "Volume detached for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-metrics-certs-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.966525 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba50945-6ce9-4a71-9f58-e436b646c779-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.966538 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.966551 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5e30fc1-acc8-448a-a9e8-490bda22e24e-ovn-northd-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.966565 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/010e4389-4656-4460-9b4d-2a7a37660252-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:14 crc kubenswrapper[4979]: I1204 12:08:14.970564 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" (UID: "000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.070857 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.075017 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_d5e30fc1-acc8-448a-a9e8-490bda22e24e/ovn-northd/0.log" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.075066 4979 generic.go:334] "Generic (PLEG): container finished" podID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" exitCode=139 Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.075131 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d5e30fc1-acc8-448a-a9e8-490bda22e24e","Type":"ContainerDied","Data":"4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a"} Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.075155 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"d5e30fc1-acc8-448a-a9e8-490bda22e24e","Type":"ContainerDied","Data":"ba4881247e039da9ed32cbe3e313560113fd6c9ec118429556cf7aecb95172e6"} Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.075162 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.075173 4979 scope.go:117] "RemoveContainer" containerID="104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.098136 4979 generic.go:334] "Generic (PLEG): container finished" podID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerID="a170026606f642d3fddc6953a178bf2cc7cbe7322dc226b38b4e679b7df1e228" exitCode=0 Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.098187 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0fad7195-d19a-48ce-ad3d-4c67b53d2974","Type":"ContainerDied","Data":"a170026606f642d3fddc6953a178bf2cc7cbe7322dc226b38b4e679b7df1e228"} Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.105719 4979 generic.go:334] "Generic (PLEG): container finished" podID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerID="ea56050492cab7388c6ef7c5a9645f7cda3ba4238e27b00f2097646a63874d12" exitCode=0 Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.106416 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novacell02b03-account-delete-hf9n8" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.106548 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48","Type":"ContainerDied","Data":"ea56050492cab7388c6ef7c5a9645f7cda3ba4238e27b00f2097646a63874d12"} Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.106955 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/novaapic904-account-delete-cvwnr" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.107591 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/barbicanee53-account-delete-gpkgc" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.107627 4979 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openstack/cinder37f5-account-delete-4wn9x" secret="" err="secret \"galera-openstack-dockercfg-7tqtd\" not found" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.352692 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.356844 4979 scope.go:117] "RemoveContainer" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377011 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-tls\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377083 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377106 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-server-conf\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377130 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-plugins\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377157 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-plugins-conf\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377190 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-pod-info\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377268 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-erlang-cookie-secret\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377353 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-confd\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377409 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-erlang-cookie\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377495 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzm4n\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-kube-api-access-jzm4n\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.377527 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data\") pod \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\" (UID: \"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.380801 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.382537 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.383592 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.383816 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.390172 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.390720 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.392478 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-kube-api-access-jzm4n" (OuterVolumeSpecName: "kube-api-access-jzm4n") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "kube-api-access-jzm4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.394640 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-pod-info" (OuterVolumeSpecName: "pod-info") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.406171 4979 scope.go:117] "RemoveContainer" containerID="104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024" Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.415438 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024\": container with ID starting with 104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024 not found: ID does not exist" containerID="104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.415691 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024"} err="failed to get container status \"104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024\": rpc error: code = NotFound desc = could not find container \"104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024\": container with ID starting with 104dd886f98da02f522719c49933ef54fc8a4122c3c164a2277df13ac9fcd024 not found: ID does not exist" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.415721 4979 scope.go:117] "RemoveContainer" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.420723 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a\": container with ID starting with 4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a not found: ID does not exist" containerID="4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.420774 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a"} err="failed to get container status \"4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a\": rpc error: code = NotFound desc = could not find container \"4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a\": container with ID starting with 4eb31826286eaf550d6b664928563cb0c506f41586154532351f83591948c19a not found: ID does not exist" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.451238 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.452005 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data" (OuterVolumeSpecName: "config-data") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.466237 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.470801 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-server-conf" (OuterVolumeSpecName: "server-conf") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.475729 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478399 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478455 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-tls\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478491 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-plugins\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478592 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-confd\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478646 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0fad7195-d19a-48ce-ad3d-4c67b53d2974-pod-info\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478811 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-server-conf\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478841 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478864 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-plugins-conf\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478903 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkr8v\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-kube-api-access-jkr8v\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478925 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-erlang-cookie\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.478959 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0fad7195-d19a-48ce-ad3d-4c67b53d2974-erlang-cookie-secret\") pod \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\" (UID: \"0fad7195-d19a-48ce-ad3d-4c67b53d2974\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479316 4979 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479337 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479347 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzm4n\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-kube-api-access-jzm4n\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479357 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479366 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479384 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479393 4979 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479402 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479411 4979 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.479420 4979 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.488920 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-kube-api-access-jkr8v" (OuterVolumeSpecName: "kube-api-access-jkr8v") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "kube-api-access-jkr8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.489288 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.489515 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.493046 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.494608 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.495557 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.496663 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fad7195-d19a-48ce-ad3d-4c67b53d2974-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.503354 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.511539 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0fad7195-d19a-48ce-ad3d-4c67b53d2974-pod-info" (OuterVolumeSpecName: "pod-info") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.524658 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.524971 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.532698 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" (UID: "0e592d2c-bb50-44b9-bd85-fdf94ee1ac48"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.533165 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data" (OuterVolumeSpecName: "config-data") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.548669 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.564896 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.571437 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582497 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582865 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582889 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582901 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582912 4979 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0fad7195-d19a-48ce-ad3d-4c67b53d2974-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582923 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582957 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582969 4979 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582981 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkr8v\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-kube-api-access-jkr8v\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.582992 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.583002 4979 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0fad7195-d19a-48ce-ad3d-4c67b53d2974-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.583011 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.584398 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-server-conf" (OuterVolumeSpecName: "server-conf") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.592586 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-proxy-db585f97f-d5777"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.627073 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.635612 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0fad7195-d19a-48ce-ad3d-4c67b53d2974" (UID: "0fad7195-d19a-48ce-ad3d-4c67b53d2974"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.635810 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-proxy-db585f97f-d5777"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.652439 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.661091 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5667db5f8d-q4ffn"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.667840 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5667db5f8d-q4ffn"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.675262 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/memcached-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.685374 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/memcached-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686104 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-kolla-config\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686210 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-operator-scripts\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686412 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mysql-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686576 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-galera-tls-certs\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686693 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-combined-ca-bundle\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686796 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-config-data-default\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686909 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jsc2q\" (UniqueName: \"kubernetes.io/projected/7130263f-15db-4a87-a078-9c346739eced-kube-api-access-jsc2q\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.686973 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-kolla-config" (OuterVolumeSpecName: "kolla-config") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "kolla-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.687126 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7130263f-15db-4a87-a078-9c346739eced-config-data-generated\") pod \"7130263f-15db-4a87-a078-9c346739eced\" (UID: \"7130263f-15db-4a87-a078-9c346739eced\") " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.687654 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0fad7195-d19a-48ce-ad3d-4c67b53d2974-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.687747 4979 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0fad7195-d19a-48ce-ad3d-4c67b53d2974-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.687837 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.687911 4979 reconciler_common.go:293] "Volume detached for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-kolla-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.687832 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.689584 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7130263f-15db-4a87-a078-9c346739eced-config-data-generated" (OuterVolumeSpecName: "config-data-generated") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "config-data-generated". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.689668 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.689713 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:08:17.689697462 +0000 UTC m=+1521.963993256 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.689721 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-config-data-default" (OuterVolumeSpecName: "config-data-default") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "config-data-default". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.691464 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.695355 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "mysql-db") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.697604 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.704362 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.706775 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7130263f-15db-4a87-a078-9c346739eced-kube-api-access-jsc2q" (OuterVolumeSpecName: "kube-api-access-jsc2q") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "kube-api-access-jsc2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.715483 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.726488 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.740216 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.752850 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.761266 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-57cfcdd946-gb75v"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.764796 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-galera-tls-certs" (OuterVolumeSpecName: "galera-tls-certs") pod "7130263f-15db-4a87-a078-9c346739eced" (UID: "7130263f-15db-4a87-a078-9c346739eced"). InnerVolumeSpecName "galera-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.770612 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-57cfcdd946-gb75v"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.780367 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789371 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jsc2q\" (UniqueName: \"kubernetes.io/projected/7130263f-15db-4a87-a078-9c346739eced-kube-api-access-jsc2q\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789411 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/7130263f-15db-4a87-a078-9c346739eced-config-data-generated\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789421 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789444 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789456 4979 reconciler_common.go:293] "Volume detached for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-galera-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789464 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7130263f-15db-4a87-a078-9c346739eced-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.789494 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/7130263f-15db-4a87-a078-9c346739eced-config-data-default\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.789986 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.790120 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:08:17.790098835 +0000 UTC m=+1522.064394699 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.790393 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.790445 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:17.790431954 +0000 UTC m=+1522.064727828 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.790516 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: E1204 12:08:15.790577 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:17.790556007 +0000 UTC m=+1522.064851811 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.793677 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.809148 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.810414 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.811117 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 12:08:15 crc kubenswrapper[4979]: I1204 12:08:15.895861 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.146805 4979 generic.go:334] "Generic (PLEG): container finished" podID="7130263f-15db-4a87-a078-9c346739eced" containerID="1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923" exitCode=0 Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.147158 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7130263f-15db-4a87-a078-9c346739eced","Type":"ContainerDied","Data":"1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.147186 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"7130263f-15db-4a87-a078-9c346739eced","Type":"ContainerDied","Data":"ac4aaca30cc1606e8c5bf93c112be1c940e01ff1a56fdd42fc23c84df674475b"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.147211 4979 scope.go:117] "RemoveContainer" containerID="1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.147546 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.154114 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.154399 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0fad7195-d19a-48ce-ad3d-4c67b53d2974","Type":"ContainerDied","Data":"35f166aeb381070445583e255246ccc0b4489934f3c5c7566f44d3721006b0bb"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.177419 4979 generic.go:334] "Generic (PLEG): container finished" podID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerID="c4914f15e8bbbd431d76efb3ff6d2212902c6e00ddaf95b123156178bfa53c1c" exitCode=0 Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.177480 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerDied","Data":"c4914f15e8bbbd431d76efb3ff6d2212902c6e00ddaf95b123156178bfa53c1c"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.184011 4979 generic.go:334] "Generic (PLEG): container finished" podID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerID="15976ba28c5c9feffdf2219916bb384ba02e85b8e1455eb217e64c14c2a43e76" exitCode=0 Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.184164 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b84847487-r8c7n" event={"ID":"869a1863-ce1d-4acd-9d50-6987b60e05f0","Type":"ContainerDied","Data":"15976ba28c5c9feffdf2219916bb384ba02e85b8e1455eb217e64c14c2a43e76"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.190764 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"0e592d2c-bb50-44b9-bd85-fdf94ee1ac48","Type":"ContainerDied","Data":"8325f2ca802d79b896407aaa3d575bc1bb498938c81b1d45cb08ee25449c8f74"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.190866 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.205769 4979 generic.go:334] "Generic (PLEG): container finished" podID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerID="b68d477aa81babb95a8e72c26e66dc3b949d75be0432ca0d51c1a917acf6f6e3" exitCode=0 Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.220545 4979 scope.go:117] "RemoveContainer" containerID="ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.223159 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" path="/var/lib/kubelet/pods/000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.223887 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="010e4389-4656-4460-9b4d-2a7a37660252" path="/var/lib/kubelet/pods/010e4389-4656-4460-9b4d-2a7a37660252/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.224975 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" path="/var/lib/kubelet/pods/122bb56e-b8fc-4822-b6f5-6c91122ff4ee/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.226227 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13394b46-eca0-4f42-b3e7-88e20f0bb590" path="/var/lib/kubelet/pods/13394b46-eca0-4f42-b3e7-88e20f0bb590/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.226883 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" path="/var/lib/kubelet/pods/1ba50945-6ce9-4a71-9f58-e436b646c779/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.227640 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" path="/var/lib/kubelet/pods/2853b847-4bf0-4d0d-af7b-ca8bc87a9c99/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.231380 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" path="/var/lib/kubelet/pods/3238d4e5-e968-4d66-87d6-7397f48e9714/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.232493 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" path="/var/lib/kubelet/pods/3cd5e649-996b-4558-b75d-b4cb3dfe2d4c/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.234919 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" path="/var/lib/kubelet/pods/4ad703ff-c521-40f1-9fa4-09fca7726079/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.236155 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" path="/var/lib/kubelet/pods/c472e576-897a-4ea6-912a-2d87b2f67077/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.237039 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" path="/var/lib/kubelet/pods/d3791d23-03e1-4c8d-b44e-738eff9222d0/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.238283 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" path="/var/lib/kubelet/pods/d5e30fc1-acc8-448a-a9e8-490bda22e24e/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.239277 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" path="/var/lib/kubelet/pods/e2d24f5a-ae53-44e3-bc4c-c56813f0f141/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.240026 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" path="/var/lib/kubelet/pods/ee632468-ca6f-48b2-abab-9cd939aafc00/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.241062 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f763e763-10df-408d-9452-84634d47b455" path="/var/lib/kubelet/pods/f763e763-10df-408d-9452-84634d47b455/volumes" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.242125 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" event={"ID":"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a","Type":"ContainerDied","Data":"b68d477aa81babb95a8e72c26e66dc3b949d75be0432ca0d51c1a917acf6f6e3"} Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.242164 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.256073 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.425955 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.430775 4979 scope.go:117] "RemoveContainer" containerID="1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923" Dec 04 12:08:16 crc kubenswrapper[4979]: E1204 12:08:16.434198 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923\": container with ID starting with 1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923 not found: ID does not exist" containerID="1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.434288 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923"} err="failed to get container status \"1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923\": rpc error: code = NotFound desc = could not find container \"1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923\": container with ID starting with 1d4627322b991fe4849199eb6bea59a126a218bed2ae4659b4b1b2a960612923 not found: ID does not exist" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.434349 4979 scope.go:117] "RemoveContainer" containerID="ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8" Dec 04 12:08:16 crc kubenswrapper[4979]: E1204 12:08:16.437637 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8\": container with ID starting with ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8 not found: ID does not exist" containerID="ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.437729 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8"} err="failed to get container status \"ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8\": rpc error: code = NotFound desc = could not find container \"ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8\": container with ID starting with ebfd028d129e3092b9b0ba73152efa3b82aaf1bb86bab3e0bcdf57d33bbae9e8 not found: ID does not exist" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.437771 4979 scope.go:117] "RemoveContainer" containerID="a170026606f642d3fddc6953a178bf2cc7cbe7322dc226b38b4e679b7df1e228" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.442801 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.452541 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.482951 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.501005 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.502269 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-b2dt6"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.518936 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-b2dt6"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.527361 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-2ebe-account-create-update-tngwj"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.536869 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance2ebe-account-delete-wzsfb"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.539701 4979 scope.go:117] "RemoveContainer" containerID="ddc7fe209b64e40b2ff31581c3560cda53976230e0b9d95e54824b9db34296d8" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.542557 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.547235 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance2ebe-account-delete-wzsfb"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.553981 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.566003 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-2ebe-account-create-update-tngwj"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.596731 4979 scope.go:117] "RemoveContainer" containerID="ea56050492cab7388c6ef7c5a9645f7cda3ba4238e27b00f2097646a63874d12" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.612907 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-combined-ca-bundle\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613014 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data-custom\") pod \"869a1863-ce1d-4acd-9d50-6987b60e05f0\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613097 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2952g\" (UniqueName: \"kubernetes.io/projected/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-kube-api-access-2952g\") pod \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613139 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-logs\") pod \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613202 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a1863-ce1d-4acd-9d50-6987b60e05f0-logs\") pod \"869a1863-ce1d-4acd-9d50-6987b60e05f0\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613226 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-combined-ca-bundle\") pod \"869a1863-ce1d-4acd-9d50-6987b60e05f0\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613318 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-scripts\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613361 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-config-data\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613436 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-run-httpd\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613459 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-combined-ca-bundle\") pod \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613480 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data-custom\") pod \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613515 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-sg-core-conf-yaml\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613536 4979 scope.go:117] "RemoveContainer" containerID="23f122939761be5f1b8c15a337a24423951f7eb2b8b6183e7f17f9e0cc30c1e6" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613549 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-ceilometer-tls-certs\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613675 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data\") pod \"869a1863-ce1d-4acd-9d50-6987b60e05f0\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613742 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/869a1863-ce1d-4acd-9d50-6987b60e05f0-kube-api-access-9kbvg\") pod \"869a1863-ce1d-4acd-9d50-6987b60e05f0\" (UID: \"869a1863-ce1d-4acd-9d50-6987b60e05f0\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613782 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-log-httpd\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613804 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data\") pod \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\" (UID: \"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.613830 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hph5n\" (UniqueName: \"kubernetes.io/projected/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-kube-api-access-hph5n\") pod \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\" (UID: \"d171ca4c-bad7-4deb-b2d2-fd6691578bb7\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.614681 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.616121 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-logs" (OuterVolumeSpecName: "logs") pod "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" (UID: "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.617633 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/869a1863-ce1d-4acd-9d50-6987b60e05f0-logs" (OuterVolumeSpecName: "logs") pod "869a1863-ce1d-4acd-9d50-6987b60e05f0" (UID: "869a1863-ce1d-4acd-9d50-6987b60e05f0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.621937 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.621969 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.621979 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/869a1863-ce1d-4acd-9d50-6987b60e05f0-logs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.623173 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.623885 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-scripts" (OuterVolumeSpecName: "scripts") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.623923 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "869a1863-ce1d-4acd-9d50-6987b60e05f0" (UID: "869a1863-ce1d-4acd-9d50-6987b60e05f0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.631612 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-kube-api-access-hph5n" (OuterVolumeSpecName: "kube-api-access-hph5n") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "kube-api-access-hph5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.633608 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" (UID: "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.633698 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-kube-api-access-2952g" (OuterVolumeSpecName: "kube-api-access-2952g") pod "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" (UID: "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a"). InnerVolumeSpecName "kube-api-access-2952g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.633789 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869a1863-ce1d-4acd-9d50-6987b60e05f0-kube-api-access-9kbvg" (OuterVolumeSpecName: "kube-api-access-9kbvg") pod "869a1863-ce1d-4acd-9d50-6987b60e05f0" (UID: "869a1863-ce1d-4acd-9d50-6987b60e05f0"). InnerVolumeSpecName "kube-api-access-9kbvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.662644 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.663730 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-wl8cj"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.678911 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" (UID: "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.679434 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-wl8cj"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.686611 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.688000 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-375f-account-create-update-bh966"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.689936 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data" (OuterVolumeSpecName: "config-data") pod "869a1863-ce1d-4acd-9d50-6987b60e05f0" (UID: "869a1863-ce1d-4acd-9d50-6987b60e05f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.690828 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data" (OuterVolumeSpecName: "config-data") pod "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" (UID: "f8589c26-b1f1-48b2-a0cd-5b1bdc62336a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.697388 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement375f-account-delete-28b5l"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.709361 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement375f-account-delete-28b5l"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.718470 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "869a1863-ce1d-4acd-9d50-6987b60e05f0" (UID: "869a1863-ce1d-4acd-9d50-6987b60e05f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.719243 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-375f-account-create-update-bh966"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723530 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723558 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hph5n\" (UniqueName: \"kubernetes.io/projected/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-kube-api-access-hph5n\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723568 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723577 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2952g\" (UniqueName: \"kubernetes.io/projected/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-kube-api-access-2952g\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723585 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723594 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723606 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723617 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723626 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723637 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723647 4979 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723658 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/869a1863-ce1d-4acd-9d50-6987b60e05f0-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.723668 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/869a1863-ce1d-4acd-9d50-6987b60e05f0-kube-api-access-9kbvg\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.732419 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.745328 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-config-data" (OuterVolumeSpecName: "config-data") pod "d171ca4c-bad7-4deb-b2d2-fd6691578bb7" (UID: "d171ca4c-bad7-4deb-b2d2-fd6691578bb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.793062 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.831569 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data\") pod \"77ca261e-57dc-4b57-a80b-e7def9b113d5\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.831772 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7b7x\" (UniqueName: \"kubernetes.io/projected/77ca261e-57dc-4b57-a80b-e7def9b113d5-kube-api-access-c7b7x\") pod \"77ca261e-57dc-4b57-a80b-e7def9b113d5\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.831812 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-combined-ca-bundle\") pod \"77ca261e-57dc-4b57-a80b-e7def9b113d5\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.832195 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.832214 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d171ca4c-bad7-4deb-b2d2-fd6691578bb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.835243 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77ca261e-57dc-4b57-a80b-e7def9b113d5-kube-api-access-c7b7x" (OuterVolumeSpecName: "kube-api-access-c7b7x") pod "77ca261e-57dc-4b57-a80b-e7def9b113d5" (UID: "77ca261e-57dc-4b57-a80b-e7def9b113d5"). InnerVolumeSpecName "kube-api-access-c7b7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: E1204 12:08:16.861707 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data podName:77ca261e-57dc-4b57-a80b-e7def9b113d5 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:17.361677184 +0000 UTC m=+1521.635972988 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data") pod "77ca261e-57dc-4b57-a80b-e7def9b113d5" (UID: "77ca261e-57dc-4b57-a80b-e7def9b113d5") : error deleting /var/lib/kubelet/pods/77ca261e-57dc-4b57-a80b-e7def9b113d5/volume-subpaths: remove /var/lib/kubelet/pods/77ca261e-57dc-4b57-a80b-e7def9b113d5/volume-subpaths: no such file or directory Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.865135 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "77ca261e-57dc-4b57-a80b-e7def9b113d5" (UID: "77ca261e-57dc-4b57-a80b-e7def9b113d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.914868 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-swzf4"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.935393 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7b7x\" (UniqueName: \"kubernetes.io/projected/77ca261e-57dc-4b57-a80b-e7def9b113d5-kube-api-access-c7b7x\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.935421 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.946468 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-swzf4"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.964084 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder37f5-account-delete-4wn9x"] Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.964319 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder37f5-account-delete-4wn9x" podUID="1341b825-bf58-4a12-af3e-40411a861e8f" containerName="mariadb-account-delete" containerID="cri-o://02302b47b9f3434931fdab45d8a9de412215496601beedc94b30eaf534866d88" gracePeriod=30 Dec 04 12:08:16 crc kubenswrapper[4979]: I1204 12:08:16.979965 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-37f5-account-create-update-m9f9n"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:16.998566 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-37f5-account-create-update-m9f9n"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.093890 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-b4thg"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.109928 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-b4thg"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.124201 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-47fe-account-create-update-dpg62"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.138742 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron47fe-account-delete-kqc4q"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.157895 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-47fe-account-create-update-dpg62"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.165292 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.167373 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron47fe-account-delete-kqc4q"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.186035 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-znf8s"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.198541 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-znf8s"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.200992 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.226537 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanee53-account-delete-gpkgc"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.226767 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbicanee53-account-delete-gpkgc" podUID="a889a605-1dbe-4a0c-a91a-3a9d2153543a" containerName="mariadb-account-delete" containerID="cri-o://3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a" gracePeriod=30 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.242728 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sk8k\" (UniqueName: \"kubernetes.io/projected/983663bf-7439-4648-8237-1174fe609970-kube-api-access-6sk8k\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243023 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-credential-keys\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243092 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-internal-tls-certs\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243194 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-config-data\") pod \"bf314b53-497e-47cd-adce-c905fff3a5b7\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243274 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptrdf\" (UniqueName: \"kubernetes.io/projected/bf314b53-497e-47cd-adce-c905fff3a5b7-kube-api-access-ptrdf\") pod \"bf314b53-497e-47cd-adce-c905fff3a5b7\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243399 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-public-tls-certs\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243491 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-fernet-keys\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243584 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-scripts\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243873 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-config-data\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.243953 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-combined-ca-bundle\") pod \"bf314b53-497e-47cd-adce-c905fff3a5b7\" (UID: \"bf314b53-497e-47cd-adce-c905fff3a5b7\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.244085 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-combined-ca-bundle\") pod \"983663bf-7439-4648-8237-1174fe609970\" (UID: \"983663bf-7439-4648-8237-1174fe609970\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.244711 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-ee53-account-create-update-qlklb"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.248766 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-scripts" (OuterVolumeSpecName: "scripts") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.248961 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/983663bf-7439-4648-8237-1174fe609970-kube-api-access-6sk8k" (OuterVolumeSpecName: "kube-api-access-6sk8k") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "kube-api-access-6sk8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.249523 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.249758 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.250501 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf314b53-497e-47cd-adce-c905fff3a5b7-kube-api-access-ptrdf" (OuterVolumeSpecName: "kube-api-access-ptrdf") pod "bf314b53-497e-47cd-adce-c905fff3a5b7" (UID: "bf314b53-497e-47cd-adce-c905fff3a5b7"). InnerVolumeSpecName "kube-api-access-ptrdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.263128 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d171ca4c-bad7-4deb-b2d2-fd6691578bb7","Type":"ContainerDied","Data":"c6962f8310b722640e160e37844cd071c65e0d294c02e5b14021d5f63c753d4b"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.263211 4979 scope.go:117] "RemoveContainer" containerID="c1c754fe678fbfa03811705e43d6a8012c1cb6310eb4b53a2378db11a7f2248c" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.263462 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.270830 4979 generic.go:334] "Generic (PLEG): container finished" podID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerID="225feb849cbf0c27ccb08142f91bf2d2a45a52845a9b008d7dd3f641eb0c0175" exitCode=0 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.270882 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698495cf57-xzqk7" event={"ID":"d5d21f75-aea0-4306-843d-c3eb782cdd78","Type":"ContainerDied","Data":"225feb849cbf0c27ccb08142f91bf2d2a45a52845a9b008d7dd3f641eb0c0175"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.273607 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-config-data" (OuterVolumeSpecName: "config-data") pod "bf314b53-497e-47cd-adce-c905fff3a5b7" (UID: "bf314b53-497e-47cd-adce-c905fff3a5b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.273712 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7b84847487-r8c7n" event={"ID":"869a1863-ce1d-4acd-9d50-6987b60e05f0","Type":"ContainerDied","Data":"209376b52f61c881989d79e3c6add103d70f6c95015802fac48f8e7f2772b22b"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.273774 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7b84847487-r8c7n" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.279033 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-config-data" (OuterVolumeSpecName: "config-data") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.279423 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-ee53-account-create-update-qlklb"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.297754 4979 generic.go:334] "Generic (PLEG): container finished" podID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" exitCode=0 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.297821 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bf314b53-497e-47cd-adce-c905fff3a5b7","Type":"ContainerDied","Data":"8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.297847 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"bf314b53-497e-47cd-adce-c905fff3a5b7","Type":"ContainerDied","Data":"5b3488c9494646b6d351ac4a684d1daf4d8f4e67ae844650672f0bc8f67b7be9"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.298307 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.298463 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.304484 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.304756 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6bf4748b96-fzr6d" event={"ID":"f8589c26-b1f1-48b2-a0cd-5b1bdc62336a","Type":"ContainerDied","Data":"e9e0d76af00c43a82906c62c6a0428bf37ccb44e1b696e17730d74ec00a096d2"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.310258 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf314b53-497e-47cd-adce-c905fff3a5b7" (UID: "bf314b53-497e-47cd-adce-c905fff3a5b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.311193 4979 generic.go:334] "Generic (PLEG): container finished" podID="983663bf-7439-4648-8237-1174fe609970" containerID="91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1" exitCode=0 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.311246 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c66cf746c-gkxxt" event={"ID":"983663bf-7439-4648-8237-1174fe609970","Type":"ContainerDied","Data":"91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.311268 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7c66cf746c-gkxxt" event={"ID":"983663bf-7439-4648-8237-1174fe609970","Type":"ContainerDied","Data":"6913d7234ddba603f928d309b2145b558a0fdda4ad8c35b0e9100d92e34e5a13"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.311408 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7c66cf746c-gkxxt" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.314926 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.315515 4979 generic.go:334] "Generic (PLEG): container finished" podID="77ca261e-57dc-4b57-a80b-e7def9b113d5" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" exitCode=0 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.315595 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77ca261e-57dc-4b57-a80b-e7def9b113d5","Type":"ContainerDied","Data":"f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.315660 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"77ca261e-57dc-4b57-a80b-e7def9b113d5","Type":"ContainerDied","Data":"350234e9448f97a3bc3e3fa4d40d902c428f14deb9688f82dadaa360f8fb5930"} Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.315729 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.346449 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "983663bf-7439-4648-8237-1174fe609970" (UID: "983663bf-7439-4648-8237-1174fe609970"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348038 4979 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348083 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348095 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348107 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348121 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348132 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sk8k\" (UniqueName: \"kubernetes.io/projected/983663bf-7439-4648-8237-1174fe609970-kube-api-access-6sk8k\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348144 4979 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348155 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348167 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf314b53-497e-47cd-adce-c905fff3a5b7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348178 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptrdf\" (UniqueName: \"kubernetes.io/projected/bf314b53-497e-47cd-adce-c905fff3a5b7-kube-api-access-ptrdf\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.348190 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/983663bf-7439-4648-8237-1174fe609970-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.378412 4979 scope.go:117] "RemoveContainer" containerID="488ba15e74f2c14428d32f7d8b8d5f64b28117aabdfcb4b764a8f0e0a1923d0f" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.397825 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-worker-7b84847487-r8c7n"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.411014 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-worker-7b84847487-r8c7n"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.416420 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.418095 4979 scope.go:117] "RemoveContainer" containerID="c4914f15e8bbbd431d76efb3ff6d2212902c6e00ddaf95b123156178bfa53c1c" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.424993 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.431246 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-keystone-listener-6bf4748b96-fzr6d"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.442156 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-keystone-listener-6bf4748b96-fzr6d"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.449830 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data\") pod \"77ca261e-57dc-4b57-a80b-e7def9b113d5\" (UID: \"77ca261e-57dc-4b57-a80b-e7def9b113d5\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.461475 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data" (OuterVolumeSpecName: "config-data") pod "77ca261e-57dc-4b57-a80b-e7def9b113d5" (UID: "77ca261e-57dc-4b57-a80b-e7def9b113d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.467781 4979 scope.go:117] "RemoveContainer" containerID="2d030700ed1d0dceefff7c85f3b95944594dfdf24ef4f4fa9ee05c4e4c775576" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.475229 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-zb6xt"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.475271 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.475527 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": dial tcp 10.217.0.203:8775: i/o timeout" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.500086 4979 scope.go:117] "RemoveContainer" containerID="15976ba28c5c9feffdf2219916bb384ba02e85b8e1455eb217e64c14c2a43e76" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.504889 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-zb6xt"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.533474 4979 scope.go:117] "RemoveContainer" containerID="de14bd59b0b06af679f770868d3c98798bbffcb02c92b7336ab4b61c2bc90a5b" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.550240 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-2b03-account-create-update-d44w9"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.556425 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77ca261e-57dc-4b57-a80b-e7def9b113d5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.567012 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell02b03-account-delete-hf9n8"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.567251 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novacell02b03-account-delete-hf9n8" podUID="0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" containerName="mariadb-account-delete" containerID="cri-o://7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423" gracePeriod=30 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.576375 4979 scope.go:117] "RemoveContainer" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.579002 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-2b03-account-create-update-d44w9"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.606949 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-ps247"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.610410 4979 scope.go:117] "RemoveContainer" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.610861 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d\": container with ID starting with 8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d not found: ID does not exist" containerID="8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.610906 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d"} err="failed to get container status \"8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d\": rpc error: code = NotFound desc = could not find container \"8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d\": container with ID starting with 8b139e76abf67a99be0c709687f20f500f876c46e8186df410dd92dbe7107d6d not found: ID does not exist" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.611042 4979 scope.go:117] "RemoveContainer" containerID="b68d477aa81babb95a8e72c26e66dc3b949d75be0432ca0d51c1a917acf6f6e3" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.615072 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-ps247"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.634871 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c904-account-create-update-c2vz5"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.636605 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.649094 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic904-account-delete-cvwnr"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.649333 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/novaapic904-account-delete-cvwnr" podUID="f9b386d0-b110-48a9-b6fe-377f0faeedc0" containerName="mariadb-account-delete" containerID="cri-o://3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b" gracePeriod=30 Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.657263 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c904-account-create-update-c2vz5"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.657448 4979 scope.go:117] "RemoveContainer" containerID="4a64fd8d676b9faa863cf1c62e916da023a1db555a91558e98ac315dc5777bab" Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.658959 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.662137 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.662571 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.662602 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.663232 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.665391 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.671566 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.673372 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.673504 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.681167 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.688026 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7c66cf746c-gkxxt"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.688576 4979 scope.go:117] "RemoveContainer" containerID="91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.700137 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7c66cf746c-gkxxt"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.710707 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.712206 4979 scope.go:117] "RemoveContainer" containerID="91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1" Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.712674 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1\": container with ID starting with 91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1 not found: ID does not exist" containerID="91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.712715 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1"} err="failed to get container status \"91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1\": rpc error: code = NotFound desc = could not find container \"91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1\": container with ID starting with 91293cb2f3aed31d375b666f2b27c033b881b4f38adf946a90714080c00f53f1 not found: ID does not exist" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.712742 4979 scope.go:117] "RemoveContainer" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.716329 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.747252 4979 scope.go:117] "RemoveContainer" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.747809 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27\": container with ID starting with f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27 not found: ID does not exist" containerID="f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.747844 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27"} err="failed to get container status \"f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27\": rpc error: code = NotFound desc = could not find container \"f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27\": container with ID starting with f6714fd11425d822e1c4caaf40b8fb47ca317216c3e10f71b58e07a29627ed27 not found: ID does not exist" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759434 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-internal-tls-certs\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759580 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-ovndb-tls-certs\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759671 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-httpd-config\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759701 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-combined-ca-bundle\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759738 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xthth\" (UniqueName: \"kubernetes.io/projected/d5d21f75-aea0-4306-843d-c3eb782cdd78-kube-api-access-xthth\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759775 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-config\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.759803 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-public-tls-certs\") pod \"d5d21f75-aea0-4306-843d-c3eb782cdd78\" (UID: \"d5d21f75-aea0-4306-843d-c3eb782cdd78\") " Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.760226 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.760292 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:08:21.760272336 +0000 UTC m=+1526.034568140 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.763635 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5d21f75-aea0-4306-843d-c3eb782cdd78-kube-api-access-xthth" (OuterVolumeSpecName: "kube-api-access-xthth") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "kube-api-access-xthth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.763729 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.802138 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.803713 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.804752 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.813845 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-config" (OuterVolumeSpecName: "config") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.824688 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "d5d21f75-aea0-4306-843d-c3eb782cdd78" (UID: "d5d21f75-aea0-4306-843d-c3eb782cdd78"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.850278 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.170:8776/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.860948 4979 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.860979 4979 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.860988 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.860998 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.861007 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xthth\" (UniqueName: \"kubernetes.io/projected/d5d21f75-aea0-4306-843d-c3eb782cdd78-kube-api-access-xthth\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.861017 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-config\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: I1204 12:08:17.861026 4979 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5d21f75-aea0-4306-843d-c3eb782cdd78-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.860982 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.861116 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:08:21.86109742 +0000 UTC m=+1526.135393244 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.861019 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.861448 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:21.861437449 +0000 UTC m=+1526.135733253 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.861028 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:17 crc kubenswrapper[4979]: E1204 12:08:17.861485 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:21.86147951 +0000 UTC m=+1526.135775314 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.207244 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" path="/var/lib/kubelet/pods/0e592d2c-bb50-44b9-bd85-fdf94ee1ac48/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.208240 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" path="/var/lib/kubelet/pods/0fad7195-d19a-48ce-ad3d-4c67b53d2974/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.209197 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1177173d-34cc-4f8c-9ac9-e2661449305f" path="/var/lib/kubelet/pods/1177173d-34cc-4f8c-9ac9-e2661449305f/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.209702 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="173171c8-4339-42ae-a4bc-be3cdc3420be" path="/var/lib/kubelet/pods/173171c8-4339-42ae-a4bc-be3cdc3420be/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.210163 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="295397f3-f237-491c-82d0-780fe406c9e1" path="/var/lib/kubelet/pods/295397f3-f237-491c-82d0-780fe406c9e1/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.210641 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29b7139c-f1fc-4637-8288-053f85ab6202" path="/var/lib/kubelet/pods/29b7139c-f1fc-4637-8288-053f85ab6202/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.211484 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" path="/var/lib/kubelet/pods/2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.211949 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dd2a291-3cb5-43ab-8f6c-c7b9517f3455" path="/var/lib/kubelet/pods/4dd2a291-3cb5-43ab-8f6c-c7b9517f3455/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.212408 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="562eca34-0bf6-4bc9-a2f7-9bf2348b78c1" path="/var/lib/kubelet/pods/562eca34-0bf6-4bc9-a2f7-9bf2348b78c1/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.213415 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7130263f-15db-4a87-a078-9c346739eced" path="/var/lib/kubelet/pods/7130263f-15db-4a87-a078-9c346739eced/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.214005 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77ca261e-57dc-4b57-a80b-e7def9b113d5" path="/var/lib/kubelet/pods/77ca261e-57dc-4b57-a80b-e7def9b113d5/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.214799 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" path="/var/lib/kubelet/pods/869a1863-ce1d-4acd-9d50-6987b60e05f0/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.216368 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91e3fea1-7f8b-454c-9033-75cd30247761" path="/var/lib/kubelet/pods/91e3fea1-7f8b-454c-9033-75cd30247761/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.217272 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="983663bf-7439-4648-8237-1174fe609970" path="/var/lib/kubelet/pods/983663bf-7439-4648-8237-1174fe609970/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.217912 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7eb59f9-ff28-4260-834c-dffefdc7db5e" path="/var/lib/kubelet/pods/a7eb59f9-ff28-4260-834c-dffefdc7db5e/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.219188 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad08b8b9-6e6c-40f3-9cf4-bb5165229869" path="/var/lib/kubelet/pods/ad08b8b9-6e6c-40f3-9cf4-bb5165229869/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.220865 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7d638c-e396-403e-974b-864194d22469" path="/var/lib/kubelet/pods/ad7d638c-e396-403e-974b-864194d22469/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.221870 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03c2a80-d7a1-4eaf-b571-16a264fdc583" path="/var/lib/kubelet/pods/b03c2a80-d7a1-4eaf-b571-16a264fdc583/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.222467 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b06a4c5f-10af-4f68-a303-f9342b97172b" path="/var/lib/kubelet/pods/b06a4c5f-10af-4f68-a303-f9342b97172b/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.223402 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" path="/var/lib/kubelet/pods/bf314b53-497e-47cd-adce-c905fff3a5b7/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.223884 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" path="/var/lib/kubelet/pods/d171ca4c-bad7-4deb-b2d2-fd6691578bb7/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.225214 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e88366f6-4404-46ae-8538-a7bd3cc9bbb5" path="/var/lib/kubelet/pods/e88366f6-4404-46ae-8538-a7bd3cc9bbb5/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.225800 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec43b717-4a9b-47b0-9007-9589c7895729" path="/var/lib/kubelet/pods/ec43b717-4a9b-47b0-9007-9589c7895729/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.226279 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f250de55-be06-43c1-9366-b3575dc45061" path="/var/lib/kubelet/pods/f250de55-be06-43c1-9366-b3575dc45061/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.226991 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" path="/var/lib/kubelet/pods/f8589c26-b1f1-48b2-a0cd-5b1bdc62336a/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.228149 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac7d10c-b3d4-4fbe-8b09-939316b41b8d" path="/var/lib/kubelet/pods/fac7d10c-b3d4-4fbe-8b09-939316b41b8d/volumes" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.332257 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-698495cf57-xzqk7" event={"ID":"d5d21f75-aea0-4306-843d-c3eb782cdd78","Type":"ContainerDied","Data":"3d5f1ea774048bd6ce4fe50f30d87ddad110465fde9f7159b9ce3766eddadf93"} Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.332328 4979 scope.go:117] "RemoveContainer" containerID="4e4c287b0e57b02ab10e8858559ba9bd0849ef197cf2a81904670a23b3ad447d" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.332271 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-698495cf57-xzqk7" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.360015 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-698495cf57-xzqk7"] Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.365267 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-698495cf57-xzqk7"] Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.365477 4979 scope.go:117] "RemoveContainer" containerID="225feb849cbf0c27ccb08142f91bf2d2a45a52845a9b008d7dd3f641eb0c0175" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.377914 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-db585f97f-d5777" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-server" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": dial tcp 10.217.0.166:8080: i/o timeout" Dec 04 12:08:18 crc kubenswrapper[4979]: I1204 12:08:18.378218 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-db585f97f-d5777" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.166:8080/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 04 12:08:20 crc kubenswrapper[4979]: I1204 12:08:20.213141 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" path="/var/lib/kubelet/pods/d5d21f75-aea0-4306-843d-c3eb782cdd78/volumes" Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.831097 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.831622 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:08:29.831595003 +0000 UTC m=+1534.105890837 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.931991 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.932066 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:08:29.932048827 +0000 UTC m=+1534.206344631 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.932134 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.932159 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:29.93215016 +0000 UTC m=+1534.206445964 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.932189 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:21 crc kubenswrapper[4979]: E1204 12:08:21.932212 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:29.932204181 +0000 UTC m=+1534.206499995 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.658941 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.659352 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.659593 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.659627 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.661108 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.662530 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.663830 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:22 crc kubenswrapper[4979]: E1204 12:08:22.663872 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.659856 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.661027 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.661024 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.662747 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.662805 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.664058 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.665946 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:27 crc kubenswrapper[4979]: E1204 12:08:27.665997 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.041653 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.041739 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.041794 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.042417 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.042475 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" gracePeriod=600 Dec 04 12:08:28 crc kubenswrapper[4979]: E1204 12:08:28.173216 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.438384 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" exitCode=0 Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.438438 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf"} Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.438477 4979 scope.go:117] "RemoveContainer" containerID="fb52f2da72fff7d717ea5b0066df40b5dc5b046752472d6fe21e232348fdf595" Dec 04 12:08:28 crc kubenswrapper[4979]: I1204 12:08:28.439245 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:08:28 crc kubenswrapper[4979]: E1204 12:08:28.439798 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.844551 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.844909 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:08:45.844894617 +0000 UTC m=+1550.119190411 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.945651 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.945752 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.945779 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:08:45.945742462 +0000 UTC m=+1550.220038306 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.945822 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:45.945802274 +0000 UTC m=+1550.220098118 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.945650 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:29 crc kubenswrapper[4979]: E1204 12:08:29.946027 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:08:45.945971598 +0000 UTC m=+1550.220267442 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.659598 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.660112 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.660499 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.660560 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.660926 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.662359 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.663879 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:32 crc kubenswrapper[4979]: E1204 12:08:32.663913 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.659550 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.662462 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.662516 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.663744 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.663807 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.665447 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.667951 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Dec 04 12:08:37 crc kubenswrapper[4979]: E1204 12:08:37.667992 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-x54hj" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.283816 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x54hj_09e1a1f2-c92b-46be-835a-8ca671f26472/ovs-vswitchd/0.log" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.285054 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375145 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-run\") pod \"09e1a1f2-c92b-46be-835a-8ca671f26472\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375200 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-etc-ovs\") pod \"09e1a1f2-c92b-46be-835a-8ca671f26472\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375232 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-log\") pod \"09e1a1f2-c92b-46be-835a-8ca671f26472\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375259 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98qtc\" (UniqueName: \"kubernetes.io/projected/09e1a1f2-c92b-46be-835a-8ca671f26472-kube-api-access-98qtc\") pod \"09e1a1f2-c92b-46be-835a-8ca671f26472\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375323 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09e1a1f2-c92b-46be-835a-8ca671f26472-scripts\") pod \"09e1a1f2-c92b-46be-835a-8ca671f26472\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375315 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-run" (OuterVolumeSpecName: "var-run") pod "09e1a1f2-c92b-46be-835a-8ca671f26472" (UID: "09e1a1f2-c92b-46be-835a-8ca671f26472"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375353 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-log" (OuterVolumeSpecName: "var-log") pod "09e1a1f2-c92b-46be-835a-8ca671f26472" (UID: "09e1a1f2-c92b-46be-835a-8ca671f26472"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375325 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-etc-ovs" (OuterVolumeSpecName: "etc-ovs") pod "09e1a1f2-c92b-46be-835a-8ca671f26472" (UID: "09e1a1f2-c92b-46be-835a-8ca671f26472"). InnerVolumeSpecName "etc-ovs". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375398 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-lib\") pod \"09e1a1f2-c92b-46be-835a-8ca671f26472\" (UID: \"09e1a1f2-c92b-46be-835a-8ca671f26472\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375538 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-lib" (OuterVolumeSpecName: "var-lib") pod "09e1a1f2-c92b-46be-835a-8ca671f26472" (UID: "09e1a1f2-c92b-46be-835a-8ca671f26472"). InnerVolumeSpecName "var-lib". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375858 4979 reconciler_common.go:293] "Volume detached for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-lib\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375870 4979 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375878 4979 reconciler_common.go:293] "Volume detached for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-etc-ovs\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.375888 4979 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/09e1a1f2-c92b-46be-835a-8ca671f26472-var-log\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.376334 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09e1a1f2-c92b-46be-835a-8ca671f26472-scripts" (OuterVolumeSpecName: "scripts") pod "09e1a1f2-c92b-46be-835a-8ca671f26472" (UID: "09e1a1f2-c92b-46be-835a-8ca671f26472"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.393755 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e1a1f2-c92b-46be-835a-8ca671f26472-kube-api-access-98qtc" (OuterVolumeSpecName: "kube-api-access-98qtc") pod "09e1a1f2-c92b-46be-835a-8ca671f26472" (UID: "09e1a1f2-c92b-46be-835a-8ca671f26472"). InnerVolumeSpecName "kube-api-access-98qtc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.442257 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477228 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") pod \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477344 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gqz7\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-kube-api-access-6gqz7\") pod \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477407 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-lock\") pod \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477462 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swift\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477521 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-cache\") pod \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\" (UID: \"4eafdbc5-ec8d-4540-a82f-b628ad83fb08\") " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477833 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98qtc\" (UniqueName: \"kubernetes.io/projected/09e1a1f2-c92b-46be-835a-8ca671f26472-kube-api-access-98qtc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.477849 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/09e1a1f2-c92b-46be-835a-8ca671f26472-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.478425 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-cache" (OuterVolumeSpecName: "cache") pod "4eafdbc5-ec8d-4540-a82f-b628ad83fb08" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08"). InnerVolumeSpecName "cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.478788 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-lock" (OuterVolumeSpecName: "lock") pod "4eafdbc5-ec8d-4540-a82f-b628ad83fb08" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08"). InnerVolumeSpecName "lock". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.481066 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "swift") pod "4eafdbc5-ec8d-4540-a82f-b628ad83fb08" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.481090 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-kube-api-access-6gqz7" (OuterVolumeSpecName: "kube-api-access-6gqz7") pod "4eafdbc5-ec8d-4540-a82f-b628ad83fb08" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08"). InnerVolumeSpecName "kube-api-access-6gqz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.483817 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "4eafdbc5-ec8d-4540-a82f-b628ad83fb08" (UID: "4eafdbc5-ec8d-4540-a82f-b628ad83fb08"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.534246 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-x54hj_09e1a1f2-c92b-46be-835a-8ca671f26472/ovs-vswitchd/0.log" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.535435 4979 generic.go:334] "Generic (PLEG): container finished" podID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" exitCode=137 Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.535475 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerDied","Data":"1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e"} Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.535522 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-x54hj" event={"ID":"09e1a1f2-c92b-46be-835a-8ca671f26472","Type":"ContainerDied","Data":"15dbebed4170dd959a37049f55a4bf615ac2122ee366747d46f50e111773915c"} Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.535544 4979 scope.go:117] "RemoveContainer" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.535580 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-x54hj" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.548066 4979 generic.go:334] "Generic (PLEG): container finished" podID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerID="a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe" exitCode=137 Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.548113 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe"} Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.548143 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"4eafdbc5-ec8d-4540-a82f-b628ad83fb08","Type":"ContainerDied","Data":"2760da9775e0b47f8dd8efebeda8de2754523160f4600c33f3a61dec8e1bccaf"} Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.548165 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.567178 4979 scope.go:117] "RemoveContainer" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.571348 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-ovs-x54hj"] Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.578319 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-ovs-x54hj"] Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.579122 4979 reconciler_common.go:293] "Volume detached for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-lock\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.579246 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.579920 4979 reconciler_common.go:293] "Volume detached for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-cache\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.580014 4979 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.580126 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gqz7\" (UniqueName: \"kubernetes.io/projected/4eafdbc5-ec8d-4540-a82f-b628ad83fb08-kube-api-access-6gqz7\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.598418 4979 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.604498 4979 scope.go:117] "RemoveContainer" containerID="42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.607192 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-storage-0"] Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.613996 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-storage-0"] Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.639438 4979 scope.go:117] "RemoveContainer" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" Dec 04 12:08:38 crc kubenswrapper[4979]: E1204 12:08:38.640327 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e\": container with ID starting with 1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e not found: ID does not exist" containerID="1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.640371 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e"} err="failed to get container status \"1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e\": rpc error: code = NotFound desc = could not find container \"1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e\": container with ID starting with 1e58a73afe8b2076140de6e159b6a19a6607315ec4dba3bcaa8d1cfce1314c0e not found: ID does not exist" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.640397 4979 scope.go:117] "RemoveContainer" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" Dec 04 12:08:38 crc kubenswrapper[4979]: E1204 12:08:38.640828 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c\": container with ID starting with 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c not found: ID does not exist" containerID="64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.640869 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c"} err="failed to get container status \"64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c\": rpc error: code = NotFound desc = could not find container \"64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c\": container with ID starting with 64669cae4183a2d1ba1711ce4a1bd730b900d92c5894a337e76e6240befa493c not found: ID does not exist" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.640889 4979 scope.go:117] "RemoveContainer" containerID="42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a" Dec 04 12:08:38 crc kubenswrapper[4979]: E1204 12:08:38.641234 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a\": container with ID starting with 42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a not found: ID does not exist" containerID="42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.641255 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a"} err="failed to get container status \"42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a\": rpc error: code = NotFound desc = could not find container \"42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a\": container with ID starting with 42446c0358ad9ebbdae1277e203a609500041cf79c5c05ec9a4abea1e3e1711a not found: ID does not exist" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.641266 4979 scope.go:117] "RemoveContainer" containerID="a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.659382 4979 scope.go:117] "RemoveContainer" containerID="1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.677782 4979 scope.go:117] "RemoveContainer" containerID="fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.684372 4979 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.697129 4979 scope.go:117] "RemoveContainer" containerID="edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.716396 4979 scope.go:117] "RemoveContainer" containerID="fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.737693 4979 scope.go:117] "RemoveContainer" containerID="58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.764498 4979 scope.go:117] "RemoveContainer" containerID="f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.785577 4979 scope.go:117] "RemoveContainer" containerID="4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.805068 4979 scope.go:117] "RemoveContainer" containerID="89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.824591 4979 scope.go:117] "RemoveContainer" containerID="f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.870095 4979 scope.go:117] "RemoveContainer" containerID="4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.898293 4979 scope.go:117] "RemoveContainer" containerID="f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.955935 4979 scope.go:117] "RemoveContainer" containerID="d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238" Dec 04 12:08:38 crc kubenswrapper[4979]: I1204 12:08:38.988321 4979 scope.go:117] "RemoveContainer" containerID="37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.021749 4979 scope.go:117] "RemoveContainer" containerID="69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.047789 4979 scope.go:117] "RemoveContainer" containerID="a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.048273 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe\": container with ID starting with a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe not found: ID does not exist" containerID="a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.048355 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe"} err="failed to get container status \"a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe\": rpc error: code = NotFound desc = could not find container \"a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe\": container with ID starting with a84b3c46c5089ad38fea853ab85509cd705eec9bcb48bf131056c429fd0f45fe not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.048378 4979 scope.go:117] "RemoveContainer" containerID="1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.048800 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509\": container with ID starting with 1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509 not found: ID does not exist" containerID="1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.048844 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509"} err="failed to get container status \"1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509\": rpc error: code = NotFound desc = could not find container \"1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509\": container with ID starting with 1171beb15ebbc113472085b68eb3313451f8836f97d38cc54b1651c6f745c509 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.048882 4979 scope.go:117] "RemoveContainer" containerID="fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.049199 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957\": container with ID starting with fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957 not found: ID does not exist" containerID="fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.049225 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957"} err="failed to get container status \"fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957\": rpc error: code = NotFound desc = could not find container \"fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957\": container with ID starting with fdd0864cf11bfcfaf4a827ebcc3616337071c75468bb015d2e2cc04c02521957 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.049241 4979 scope.go:117] "RemoveContainer" containerID="edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.049597 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55\": container with ID starting with edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55 not found: ID does not exist" containerID="edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.049633 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55"} err="failed to get container status \"edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55\": rpc error: code = NotFound desc = could not find container \"edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55\": container with ID starting with edd8371b60dbb3ca51bca9bfef83cf830da893b445888e5158663bc55d596c55 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.049653 4979 scope.go:117] "RemoveContainer" containerID="fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.049973 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5\": container with ID starting with fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5 not found: ID does not exist" containerID="fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.049999 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5"} err="failed to get container status \"fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5\": rpc error: code = NotFound desc = could not find container \"fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5\": container with ID starting with fda19412a9b114782588d11035144dedbed595ba72754ca2af218c24aa516cd5 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050016 4979 scope.go:117] "RemoveContainer" containerID="58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.050231 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e\": container with ID starting with 58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e not found: ID does not exist" containerID="58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050250 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e"} err="failed to get container status \"58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e\": rpc error: code = NotFound desc = could not find container \"58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e\": container with ID starting with 58eab6e5ebb1b9e1821018e471a22c6deea6ed6781072e468e95ee536c4ba46e not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050264 4979 scope.go:117] "RemoveContainer" containerID="f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.050622 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab\": container with ID starting with f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab not found: ID does not exist" containerID="f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050645 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab"} err="failed to get container status \"f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab\": rpc error: code = NotFound desc = could not find container \"f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab\": container with ID starting with f41b569598b16dbae69cc05706517c90dc1bef1da1c675fe047280829b0ad1ab not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050659 4979 scope.go:117] "RemoveContainer" containerID="4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.050893 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01\": container with ID starting with 4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01 not found: ID does not exist" containerID="4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050917 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01"} err="failed to get container status \"4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01\": rpc error: code = NotFound desc = could not find container \"4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01\": container with ID starting with 4880a7bcd1d12520c12f7ce5df16b3357a9c2573d90289c94706a3dec059bd01 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.050934 4979 scope.go:117] "RemoveContainer" containerID="89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.051144 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757\": container with ID starting with 89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757 not found: ID does not exist" containerID="89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051166 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757"} err="failed to get container status \"89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757\": rpc error: code = NotFound desc = could not find container \"89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757\": container with ID starting with 89300481cea91f3c51e7b4b120c84573b33d74a4daa9680b2b4ec1c2f43fe757 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051178 4979 scope.go:117] "RemoveContainer" containerID="f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.051472 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf\": container with ID starting with f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf not found: ID does not exist" containerID="f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051499 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf"} err="failed to get container status \"f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf\": rpc error: code = NotFound desc = could not find container \"f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf\": container with ID starting with f50a0588cb620e71d487be15712da8ac87ff187053e6b31fdc0779ad3e8bd4bf not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051513 4979 scope.go:117] "RemoveContainer" containerID="4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.051721 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946\": container with ID starting with 4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946 not found: ID does not exist" containerID="4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051741 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946"} err="failed to get container status \"4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946\": rpc error: code = NotFound desc = could not find container \"4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946\": container with ID starting with 4fe18f29ee18b30763e8d597ff81a25d11c166fefd3970f8bb6d6b73ec7da946 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051751 4979 scope.go:117] "RemoveContainer" containerID="f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.051956 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15\": container with ID starting with f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15 not found: ID does not exist" containerID="f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051979 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15"} err="failed to get container status \"f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15\": rpc error: code = NotFound desc = could not find container \"f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15\": container with ID starting with f554bb287682a369230c939b79fcba7a8e3ae7c7d80d17f9a0f469abfa86fa15 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.051993 4979 scope.go:117] "RemoveContainer" containerID="d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.052429 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238\": container with ID starting with d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238 not found: ID does not exist" containerID="d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.052543 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238"} err="failed to get container status \"d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238\": rpc error: code = NotFound desc = could not find container \"d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238\": container with ID starting with d7e8e20c228e9acb1d5aa7cfd0dc39ddc408e0b0de5038046b37a1427ef84238 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.052600 4979 scope.go:117] "RemoveContainer" containerID="37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.053012 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6\": container with ID starting with 37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6 not found: ID does not exist" containerID="37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.053043 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6"} err="failed to get container status \"37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6\": rpc error: code = NotFound desc = could not find container \"37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6\": container with ID starting with 37e43d4b1eb2a8054471a27467b807de28d8ffbf72df0543a93bef7303d35db6 not found: ID does not exist" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.053060 4979 scope.go:117] "RemoveContainer" containerID="69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254" Dec 04 12:08:39 crc kubenswrapper[4979]: E1204 12:08:39.053350 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254\": container with ID starting with 69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254 not found: ID does not exist" containerID="69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254" Dec 04 12:08:39 crc kubenswrapper[4979]: I1204 12:08:39.053392 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254"} err="failed to get container status \"69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254\": rpc error: code = NotFound desc = could not find container \"69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254\": container with ID starting with 69fd27995bb0d501e97d7e847a139a957922265155bbf2e1c4891f1dae34e254 not found: ID does not exist" Dec 04 12:08:40 crc kubenswrapper[4979]: I1204 12:08:40.210224 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" path="/var/lib/kubelet/pods/09e1a1f2-c92b-46be-835a-8ca671f26472/volumes" Dec 04 12:08:40 crc kubenswrapper[4979]: I1204 12:08:40.211876 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" path="/var/lib/kubelet/pods/4eafdbc5-ec8d-4540-a82f-b628ad83fb08/volumes" Dec 04 12:08:41 crc kubenswrapper[4979]: I1204 12:08:41.199405 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:08:41 crc kubenswrapper[4979]: E1204 12:08:41.199817 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:08:41 crc kubenswrapper[4979]: I1204 12:08:41.683440 4979 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod70791582-cb57-4564-8a5b-9953f4f22c23"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod70791582-cb57-4564-8a5b-9953f4f22c23] : Timed out while waiting for systemd to remove kubepods-besteffort-pod70791582_cb57_4564_8a5b_9953f4f22c23.slice" Dec 04 12:08:41 crc kubenswrapper[4979]: E1204 12:08:41.683492 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod70791582-cb57-4564-8a5b-9953f4f22c23] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod70791582-cb57-4564-8a5b-9953f4f22c23] : Timed out while waiting for systemd to remove kubepods-besteffort-pod70791582_cb57_4564_8a5b_9953f4f22c23.slice" pod="openstack/ovn-controller-metrics-v6652" podUID="70791582-cb57-4564-8a5b-9953f4f22c23" Dec 04 12:08:42 crc kubenswrapper[4979]: I1204 12:08:42.110128 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-v6652" Dec 04 12:08:42 crc kubenswrapper[4979]: I1204 12:08:42.130441 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-metrics-v6652"] Dec 04 12:08:42 crc kubenswrapper[4979]: I1204 12:08:42.137610 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-metrics-v6652"] Dec 04 12:08:42 crc kubenswrapper[4979]: I1204 12:08:42.207813 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70791582-cb57-4564-8a5b-9953f4f22c23" path="/var/lib/kubelet/pods/70791582-cb57-4564-8a5b-9953f4f22c23/volumes" Dec 04 12:08:43 crc kubenswrapper[4979]: I1204 12:08:43.104081 4979 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podbdd8e97d-9bc0-4e10-984b-7d92698e8bd6"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podbdd8e97d-9bc0-4e10-984b-7d92698e8bd6] : Timed out while waiting for systemd to remove kubepods-besteffort-podbdd8e97d_9bc0_4e10_984b_7d92698e8bd6.slice" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.889074 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.889440 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts podName:1341b825-bf58-4a12-af3e-40411a861e8f nodeName:}" failed. No retries permitted until 2025-12-04 12:09:17.889424264 +0000 UTC m=+1582.163720068 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts") pod "cinder37f5-account-delete-4wn9x" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f") : configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.987997 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7246"] Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988280 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988310 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988333 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988342 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988355 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="cinder-scheduler" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988363 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="cinder-scheduler" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988372 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988378 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988385 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerName="mysql-bootstrap" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988390 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerName="mysql-bootstrap" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988400 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988406 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988414 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988422 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988431 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="sg-core" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988437 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="sg-core" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988444 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="setup-container" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988449 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="setup-container" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988460 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="setup-container" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988467 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="setup-container" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988473 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988479 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988489 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-updater" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988494 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-updater" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988502 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="983663bf-7439-4648-8237-1174fe609970" containerName="keystone-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988508 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="983663bf-7439-4648-8237-1174fe609970" containerName="keystone-api" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988518 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988525 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988532 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7130263f-15db-4a87-a078-9c346739eced" containerName="mysql-bootstrap" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988537 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7130263f-15db-4a87-a078-9c346739eced" containerName="mysql-bootstrap" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988546 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988552 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-server" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988559 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988565 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988573 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="rabbitmq" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988579 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="rabbitmq" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988589 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988594 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-api" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988600 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerName="init" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988606 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerName="init" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988613 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988619 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988628 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988634 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988640 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988646 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-server" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988655 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988662 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988673 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="proxy-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988679 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="proxy-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988691 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="ovn-northd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988696 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="ovn-northd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988708 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="ovsdbserver-sb" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988714 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="ovsdbserver-sb" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988721 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988726 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988734 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70791582-cb57-4564-8a5b-9953f4f22c23" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988740 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="70791582-cb57-4564-8a5b-9953f4f22c23" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988748 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec43b717-4a9b-47b0-9007-9589c7895729" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988754 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec43b717-4a9b-47b0-9007-9589c7895729" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988765 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988771 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988781 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988787 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988794 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988800 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988809 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-updater" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988815 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-updater" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988823 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13394b46-eca0-4f42-b3e7-88e20f0bb590" containerName="kube-state-metrics" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988829 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="13394b46-eca0-4f42-b3e7-88e20f0bb590" containerName="kube-state-metrics" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988838 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988844 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988852 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988858 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988866 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="probe" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988872 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="probe" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988878 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-metadata" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988884 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-metadata" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988893 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="ovsdbserver-nb" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988899 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="ovsdbserver-nb" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988905 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988911 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988918 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988924 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988933 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988939 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988951 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f763e763-10df-408d-9452-84634d47b455" containerName="memcached" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988956 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f763e763-10df-408d-9452-84634d47b455" containerName="memcached" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988965 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988970 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-server" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988977 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988982 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.988991 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.988997 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-api" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989007 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989013 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989023 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989029 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989036 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerName="galera" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989042 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerName="galera" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989050 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989077 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989085 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989091 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-server" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989097 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server-init" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989103 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server-init" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989112 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="swift-recon-cron" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989117 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="swift-recon-cron" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989124 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989131 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.989139 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ca261e-57dc-4b57-a80b-e7def9b113d5" containerName="nova-scheduler-scheduler" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.989145 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ca261e-57dc-4b57-a80b-e7def9b113d5" containerName="nova-scheduler-scheduler" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991120 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b06a4c5f-10af-4f68-a303-f9342b97172b" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991135 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b06a4c5f-10af-4f68-a303-f9342b97172b" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991150 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="rabbitmq" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991165 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="rabbitmq" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991174 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991180 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991189 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991195 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991203 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991210 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-api" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991218 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991224 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991231 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerName="nova-cell1-conductor-conductor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991237 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerName="nova-cell1-conductor-conductor" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991244 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7130263f-15db-4a87-a078-9c346739eced" containerName="galera" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991250 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7130263f-15db-4a87-a078-9c346739eced" containerName="galera" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991259 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-reaper" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991265 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-reaper" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991275 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" containerName="nova-cell0-conductor-conductor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991281 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" containerName="nova-cell0-conductor-conductor" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991288 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-notification-agent" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991315 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-notification-agent" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991325 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991333 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991344 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="rsync" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991351 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="rsync" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991364 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991355 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991439 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991445 4979 configmap.go:193] Couldn't get configMap openstack/openstack-scripts: configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991466 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts podName:f9b386d0-b110-48a9-b6fe-377f0faeedc0 nodeName:}" failed. No retries permitted until 2025-12-04 12:09:17.991442381 +0000 UTC m=+1582.265738205 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts") pod "novaapic904-account-delete-cvwnr" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0") : configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991371 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991521 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-expirer" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991541 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-expirer" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991561 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991493 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts podName:0c3a9020-3ac9-4bf8-b481-213c5a6a41d4 nodeName:}" failed. No retries permitted until 2025-12-04 12:09:17.991477912 +0000 UTC m=+1582.265773716 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts") pod "novacell02b03-account-delete-hf9n8" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4") : configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991640 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts podName:a889a605-1dbe-4a0c-a91a-3a9d2153543a nodeName:}" failed. No retries permitted until 2025-12-04 12:09:17.991627676 +0000 UTC m=+1582.265923500 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "operator-scripts" (UniqueName: "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts") pod "barbicanee53-account-delete-gpkgc" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a") : configmap "openstack-scripts" not found Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991641 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991664 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-central-agent" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991673 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-central-agent" Dec 04 12:08:45 crc kubenswrapper[4979]: E1204 12:08:45.991688 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerName="dnsmasq-dns" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991696 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerName="dnsmasq-dns" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991886 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991898 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991907 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991921 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c472e576-897a-4ea6-912a-2d87b2f67077" containerName="nova-api-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991929 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7130263f-15db-4a87-a078-9c346739eced" containerName="galera" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991941 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf314b53-497e-47cd-adce-c905fff3a5b7" containerName="nova-cell1-conductor-conductor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991954 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991972 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="70791582-cb57-4564-8a5b-9953f4f22c23" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991982 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.991991 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-notification-agent" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992000 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992010 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovs-vswitchd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992021 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992033 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3015ecd7-aacc-41ac-897f-083c4cb5bd62" containerName="dnsmasq-dns" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992055 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="proxy-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992065 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ca261e-57dc-4b57-a80b-e7def9b113d5" containerName="nova-scheduler-scheduler" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992078 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3791d23-03e1-4c8d-b44e-738eff9222d0" containerName="galera" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992087 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992099 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992111 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-updater" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992120 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="probe" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992130 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f763e763-10df-408d-9452-84634d47b455" containerName="memcached" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992137 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac61c2b8-6dc7-41c5-a2e2-63759d5671b6" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992149 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="ovsdbserver-nb" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992158 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992168 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="000ae0aa-a3cb-4f7b-b622-fd2b1d1075dd" containerName="placement-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992176 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992183 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-metadata" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992193 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="sg-core" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992201 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992213 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992226 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="rsync" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992234 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ba50945-6ce9-4a71-9f58-e436b646c779" containerName="barbican-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992243 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3238d4e5-e968-4d66-87d6-7397f48e9714" containerName="nova-metadata-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992253 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="010e4389-4656-4460-9b4d-2a7a37660252" containerName="cinder-scheduler" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992261 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992271 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="122bb56e-b8fc-4822-b6f5-6c91122ff4ee" containerName="nova-cell0-conductor-conductor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992280 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992291 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="account-reaper" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992320 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-replicator" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992334 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-expirer" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992346 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992357 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e592d2c-bb50-44b9-bd85-fdf94ee1ac48" containerName="rabbitmq" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992368 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="swift-recon-cron" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992377 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b06a4c5f-10af-4f68-a303-f9342b97172b" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992386 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e1a1f2-c92b-46be-835a-8ca671f26472" containerName="ovsdb-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992394 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee632468-ca6f-48b2-abab-9cd939aafc00" containerName="proxy-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992406 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd5e649-996b-4558-b75d-b4cb3dfe2d4c" containerName="glance-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992416 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2853b847-4bf0-4d0d-af7b-ca8bc87a9c99" containerName="ovn-controller" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992425 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992433 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec43b717-4a9b-47b0-9007-9589c7895729" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992446 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdd8e97d-9bc0-4e10-984b-7d92698e8bd6" containerName="ovsdbserver-sb" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992459 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b2ecb0e-a2d5-482f-8f10-3b32a0a3d941" containerName="mariadb-account-delete" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992470 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5e30fc1-acc8-448a-a9e8-490bda22e24e" containerName="ovn-northd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992482 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-server" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992494 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2d24f5a-ae53-44e3-bc4c-c56813f0f141" containerName="glance-httpd" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992502 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992515 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="object-auditor" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992526 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8589c26-b1f1-48b2-a0cd-5b1bdc62336a" containerName="barbican-keystone-listener-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992534 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ad703ff-c521-40f1-9fa4-09fca7726079" containerName="cinder-api-log" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992545 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef961b3d-b927-42e5-91b9-eb966bdee861" containerName="openstack-network-exporter" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992555 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5d21f75-aea0-4306-843d-c3eb782cdd78" containerName="neutron-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992566 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d171ca4c-bad7-4deb-b2d2-fd6691578bb7" containerName="ceilometer-central-agent" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992577 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eafdbc5-ec8d-4540-a82f-b628ad83fb08" containerName="container-updater" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992585 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="983663bf-7439-4648-8237-1174fe609970" containerName="keystone-api" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992593 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="13394b46-eca0-4f42-b3e7-88e20f0bb590" containerName="kube-state-metrics" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992602 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fad7195-d19a-48ce-ad3d-4c67b53d2974" containerName="rabbitmq" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.992610 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="869a1863-ce1d-4acd-9d50-6987b60e05f0" containerName="barbican-worker" Dec 04 12:08:45 crc kubenswrapper[4979]: I1204 12:08:45.994012 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.003489 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7246"] Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.092648 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-catalog-content\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.092711 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkn8r\" (UniqueName: \"kubernetes.io/projected/221352a3-a603-4a91-bba1-d7556bb94d16-kube-api-access-mkn8r\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.092759 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-utilities\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.193926 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-catalog-content\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.194983 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkn8r\" (UniqueName: \"kubernetes.io/projected/221352a3-a603-4a91-bba1-d7556bb94d16-kube-api-access-mkn8r\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.195222 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-utilities\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.195850 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-utilities\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.194895 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-catalog-content\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.221362 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkn8r\" (UniqueName: \"kubernetes.io/projected/221352a3-a603-4a91-bba1-d7556bb94d16-kube-api-access-mkn8r\") pod \"community-operators-t7246\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.325538 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:46 crc kubenswrapper[4979]: I1204 12:08:46.834173 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7246"] Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.153929 4979 generic.go:334] "Generic (PLEG): container finished" podID="221352a3-a603-4a91-bba1-d7556bb94d16" containerID="a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65" exitCode=0 Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.154037 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7246" event={"ID":"221352a3-a603-4a91-bba1-d7556bb94d16","Type":"ContainerDied","Data":"a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65"} Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.154639 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7246" event={"ID":"221352a3-a603-4a91-bba1-d7556bb94d16","Type":"ContainerStarted","Data":"9aa5ae4955a5cc05d17e26206953e921d3f9cdd1056e02ee26289e3612e5d653"} Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.161063 4979 generic.go:334] "Generic (PLEG): container finished" podID="1341b825-bf58-4a12-af3e-40411a861e8f" containerID="02302b47b9f3434931fdab45d8a9de412215496601beedc94b30eaf534866d88" exitCode=137 Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.161106 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder37f5-account-delete-4wn9x" event={"ID":"1341b825-bf58-4a12-af3e-40411a861e8f","Type":"ContainerDied","Data":"02302b47b9f3434931fdab45d8a9de412215496601beedc94b30eaf534866d88"} Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.744064 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.786733 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.819961 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts\") pod \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.820075 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5hnp\" (UniqueName: \"kubernetes.io/projected/a889a605-1dbe-4a0c-a91a-3a9d2153543a-kube-api-access-n5hnp\") pod \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\" (UID: \"a889a605-1dbe-4a0c-a91a-3a9d2153543a\") " Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.821679 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a889a605-1dbe-4a0c-a91a-3a9d2153543a" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.826641 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a889a605-1dbe-4a0c-a91a-3a9d2153543a-kube-api-access-n5hnp" (OuterVolumeSpecName: "kube-api-access-n5hnp") pod "a889a605-1dbe-4a0c-a91a-3a9d2153543a" (UID: "a889a605-1dbe-4a0c-a91a-3a9d2153543a"). InnerVolumeSpecName "kube-api-access-n5hnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.921190 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts\") pod \"1341b825-bf58-4a12-af3e-40411a861e8f\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.921908 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1341b825-bf58-4a12-af3e-40411a861e8f" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.921291 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jzcr9\" (UniqueName: \"kubernetes.io/projected/1341b825-bf58-4a12-af3e-40411a861e8f-kube-api-access-jzcr9\") pod \"1341b825-bf58-4a12-af3e-40411a861e8f\" (UID: \"1341b825-bf58-4a12-af3e-40411a861e8f\") " Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.922400 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5hnp\" (UniqueName: \"kubernetes.io/projected/a889a605-1dbe-4a0c-a91a-3a9d2153543a-kube-api-access-n5hnp\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.922418 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1341b825-bf58-4a12-af3e-40411a861e8f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.922430 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a889a605-1dbe-4a0c-a91a-3a9d2153543a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.930258 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1341b825-bf58-4a12-af3e-40411a861e8f-kube-api-access-jzcr9" (OuterVolumeSpecName: "kube-api-access-jzcr9") pod "1341b825-bf58-4a12-af3e-40411a861e8f" (UID: "1341b825-bf58-4a12-af3e-40411a861e8f"). InnerVolumeSpecName "kube-api-access-jzcr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:47 crc kubenswrapper[4979]: I1204 12:08:47.995148 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.024250 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jzcr9\" (UniqueName: \"kubernetes.io/projected/1341b825-bf58-4a12-af3e-40411a861e8f-kube-api-access-jzcr9\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.054060 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.125107 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fh88\" (UniqueName: \"kubernetes.io/projected/f9b386d0-b110-48a9-b6fe-377f0faeedc0-kube-api-access-2fh88\") pod \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.125181 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts\") pod \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.125421 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xwsw\" (UniqueName: \"kubernetes.io/projected/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-kube-api-access-9xwsw\") pod \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\" (UID: \"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4\") " Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.125475 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts\") pod \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\" (UID: \"f9b386d0-b110-48a9-b6fe-377f0faeedc0\") " Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.125958 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.126740 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9b386d0-b110-48a9-b6fe-377f0faeedc0" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.129620 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9b386d0-b110-48a9-b6fe-377f0faeedc0-kube-api-access-2fh88" (OuterVolumeSpecName: "kube-api-access-2fh88") pod "f9b386d0-b110-48a9-b6fe-377f0faeedc0" (UID: "f9b386d0-b110-48a9-b6fe-377f0faeedc0"). InnerVolumeSpecName "kube-api-access-2fh88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.129719 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-kube-api-access-9xwsw" (OuterVolumeSpecName: "kube-api-access-9xwsw") pod "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" (UID: "0c3a9020-3ac9-4bf8-b481-213c5a6a41d4"). InnerVolumeSpecName "kube-api-access-9xwsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.172718 4979 generic.go:334] "Generic (PLEG): container finished" podID="f9b386d0-b110-48a9-b6fe-377f0faeedc0" containerID="3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b" exitCode=137 Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.172781 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic904-account-delete-cvwnr" event={"ID":"f9b386d0-b110-48a9-b6fe-377f0faeedc0","Type":"ContainerDied","Data":"3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.172810 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novaapic904-account-delete-cvwnr" event={"ID":"f9b386d0-b110-48a9-b6fe-377f0faeedc0","Type":"ContainerDied","Data":"8caf959e5d5213b30979b37f6649552a9ef3dc51101f39a59af195c3aa422b3e"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.172826 4979 scope.go:117] "RemoveContainer" containerID="3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.172925 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novaapic904-account-delete-cvwnr" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.176535 4979 generic.go:334] "Generic (PLEG): container finished" podID="0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" containerID="7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423" exitCode=137 Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.176613 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02b03-account-delete-hf9n8" event={"ID":"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4","Type":"ContainerDied","Data":"7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.176640 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/novacell02b03-account-delete-hf9n8" event={"ID":"0c3a9020-3ac9-4bf8-b481-213c5a6a41d4","Type":"ContainerDied","Data":"bdb561f2fd816a098e8a480f61594120050d3c8b7811b4f236cc21ce18301d04"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.176671 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/novacell02b03-account-delete-hf9n8" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.178260 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder37f5-account-delete-4wn9x" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.178265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder37f5-account-delete-4wn9x" event={"ID":"1341b825-bf58-4a12-af3e-40411a861e8f","Type":"ContainerDied","Data":"ae2aa6d8b077f554ae691f32a9697c1526209e0172b29c78569ae88e8a8529ba"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.179734 4979 generic.go:334] "Generic (PLEG): container finished" podID="a889a605-1dbe-4a0c-a91a-3a9d2153543a" containerID="3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a" exitCode=137 Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.179838 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanee53-account-delete-gpkgc" event={"ID":"a889a605-1dbe-4a0c-a91a-3a9d2153543a","Type":"ContainerDied","Data":"3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.179863 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbicanee53-account-delete-gpkgc" event={"ID":"a889a605-1dbe-4a0c-a91a-3a9d2153543a","Type":"ContainerDied","Data":"0f6f3736f22758f16d22c55db1b531bcbd6407c3f1d52b34edd4bd38a33ba6be"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.179919 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbicanee53-account-delete-gpkgc" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.186373 4979 generic.go:334] "Generic (PLEG): container finished" podID="221352a3-a603-4a91-bba1-d7556bb94d16" containerID="094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963" exitCode=0 Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.186418 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7246" event={"ID":"221352a3-a603-4a91-bba1-d7556bb94d16","Type":"ContainerDied","Data":"094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963"} Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.226765 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9b386d0-b110-48a9-b6fe-377f0faeedc0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.226796 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fh88\" (UniqueName: \"kubernetes.io/projected/f9b386d0-b110-48a9-b6fe-377f0faeedc0-kube-api-access-2fh88\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.226806 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.226816 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xwsw\" (UniqueName: \"kubernetes.io/projected/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4-kube-api-access-9xwsw\") on node \"crc\" DevicePath \"\"" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.303644 4979 scope.go:117] "RemoveContainer" containerID="3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b" Dec 04 12:08:48 crc kubenswrapper[4979]: E1204 12:08:48.304399 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b\": container with ID starting with 3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b not found: ID does not exist" containerID="3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.304479 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b"} err="failed to get container status \"3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b\": rpc error: code = NotFound desc = could not find container \"3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b\": container with ID starting with 3502e15115a0486b7d16d2443770455110eb6f7a1fa04baee576a9de9512967b not found: ID does not exist" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.304538 4979 scope.go:117] "RemoveContainer" containerID="7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.333930 4979 scope.go:117] "RemoveContainer" containerID="7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423" Dec 04 12:08:48 crc kubenswrapper[4979]: E1204 12:08:48.334587 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423\": container with ID starting with 7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423 not found: ID does not exist" containerID="7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.334635 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423"} err="failed to get container status \"7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423\": rpc error: code = NotFound desc = could not find container \"7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423\": container with ID starting with 7bf9d39bb762dd2d0106251f6f7c7c695faa6b77dace947c82acd4835d796423 not found: ID does not exist" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.334661 4979 scope.go:117] "RemoveContainer" containerID="02302b47b9f3434931fdab45d8a9de412215496601beedc94b30eaf534866d88" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.340785 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novaapic904-account-delete-cvwnr"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.359976 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novaapic904-account-delete-cvwnr"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.366192 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/novacell02b03-account-delete-hf9n8"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.368159 4979 scope.go:117] "RemoveContainer" containerID="3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.370813 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/novacell02b03-account-delete-hf9n8"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.375674 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder37f5-account-delete-4wn9x"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.381017 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder37f5-account-delete-4wn9x"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.386472 4979 scope.go:117] "RemoveContainer" containerID="3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a" Dec 04 12:08:48 crc kubenswrapper[4979]: E1204 12:08:48.386895 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a\": container with ID starting with 3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a not found: ID does not exist" containerID="3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.386924 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a"} err="failed to get container status \"3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a\": rpc error: code = NotFound desc = could not find container \"3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a\": container with ID starting with 3ee59b5bb032cd9d7f9b8be55aef4a291799cfb34b25e14a49b76c5f0864c91a not found: ID does not exist" Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.387377 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbicanee53-account-delete-gpkgc"] Dec 04 12:08:48 crc kubenswrapper[4979]: I1204 12:08:48.393155 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbicanee53-account-delete-gpkgc"] Dec 04 12:08:49 crc kubenswrapper[4979]: I1204 12:08:49.212918 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7246" event={"ID":"221352a3-a603-4a91-bba1-d7556bb94d16","Type":"ContainerStarted","Data":"4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9"} Dec 04 12:08:49 crc kubenswrapper[4979]: I1204 12:08:49.237727 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7246" podStartSLOduration=2.812468055 podStartE2EDuration="4.237689375s" podCreationTimestamp="2025-12-04 12:08:45 +0000 UTC" firstStartedPulling="2025-12-04 12:08:47.158326015 +0000 UTC m=+1551.432621819" lastFinishedPulling="2025-12-04 12:08:48.583547335 +0000 UTC m=+1552.857843139" observedRunningTime="2025-12-04 12:08:49.231628113 +0000 UTC m=+1553.505923927" watchObservedRunningTime="2025-12-04 12:08:49.237689375 +0000 UTC m=+1553.511985179" Dec 04 12:08:50 crc kubenswrapper[4979]: I1204 12:08:50.211787 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" path="/var/lib/kubelet/pods/0c3a9020-3ac9-4bf8-b481-213c5a6a41d4/volumes" Dec 04 12:08:50 crc kubenswrapper[4979]: I1204 12:08:50.212496 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1341b825-bf58-4a12-af3e-40411a861e8f" path="/var/lib/kubelet/pods/1341b825-bf58-4a12-af3e-40411a861e8f/volumes" Dec 04 12:08:50 crc kubenswrapper[4979]: I1204 12:08:50.213149 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a889a605-1dbe-4a0c-a91a-3a9d2153543a" path="/var/lib/kubelet/pods/a889a605-1dbe-4a0c-a91a-3a9d2153543a/volumes" Dec 04 12:08:50 crc kubenswrapper[4979]: I1204 12:08:50.213886 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9b386d0-b110-48a9-b6fe-377f0faeedc0" path="/var/lib/kubelet/pods/f9b386d0-b110-48a9-b6fe-377f0faeedc0/volumes" Dec 04 12:08:55 crc kubenswrapper[4979]: I1204 12:08:55.199450 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:08:55 crc kubenswrapper[4979]: E1204 12:08:55.200068 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.326035 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.326328 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.369032 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820029 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2g6ml"] Dec 04 12:08:56 crc kubenswrapper[4979]: E1204 12:08:56.820356 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1341b825-bf58-4a12-af3e-40411a861e8f" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820371 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1341b825-bf58-4a12-af3e-40411a861e8f" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: E1204 12:08:56.820395 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a889a605-1dbe-4a0c-a91a-3a9d2153543a" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820401 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a889a605-1dbe-4a0c-a91a-3a9d2153543a" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: E1204 12:08:56.820415 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9b386d0-b110-48a9-b6fe-377f0faeedc0" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820421 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9b386d0-b110-48a9-b6fe-377f0faeedc0" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: E1204 12:08:56.820431 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820437 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820555 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c3a9020-3ac9-4bf8-b481-213c5a6a41d4" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820566 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a889a605-1dbe-4a0c-a91a-3a9d2153543a" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820580 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9b386d0-b110-48a9-b6fe-377f0faeedc0" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.820594 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1341b825-bf58-4a12-af3e-40411a861e8f" containerName="mariadb-account-delete" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.821586 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.831125 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2g6ml"] Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.949804 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvqvk\" (UniqueName: \"kubernetes.io/projected/8aea49a0-460d-4dc2-8132-e13a34d4cced-kube-api-access-bvqvk\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.949835 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-utilities\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:56 crc kubenswrapper[4979]: I1204 12:08:56.949875 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-catalog-content\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.051229 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvqvk\" (UniqueName: \"kubernetes.io/projected/8aea49a0-460d-4dc2-8132-e13a34d4cced-kube-api-access-bvqvk\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.051594 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-utilities\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.051712 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-catalog-content\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.052233 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-catalog-content\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.052264 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-utilities\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.085124 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvqvk\" (UniqueName: \"kubernetes.io/projected/8aea49a0-460d-4dc2-8132-e13a34d4cced-kube-api-access-bvqvk\") pod \"certified-operators-2g6ml\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.142000 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.349534 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:08:57 crc kubenswrapper[4979]: I1204 12:08:57.606473 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2g6ml"] Dec 04 12:08:58 crc kubenswrapper[4979]: I1204 12:08:58.293950 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2g6ml" event={"ID":"8aea49a0-460d-4dc2-8132-e13a34d4cced","Type":"ContainerStarted","Data":"9f85ba2be91e41e7277dbbcab1d302cdda4dae2ad92b1c95d611a5fbefaf2d03"} Dec 04 12:08:59 crc kubenswrapper[4979]: I1204 12:08:59.303905 4979 generic.go:334] "Generic (PLEG): container finished" podID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerID="bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6" exitCode=0 Dec 04 12:08:59 crc kubenswrapper[4979]: I1204 12:08:59.303963 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2g6ml" event={"ID":"8aea49a0-460d-4dc2-8132-e13a34d4cced","Type":"ContainerDied","Data":"bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6"} Dec 04 12:08:59 crc kubenswrapper[4979]: I1204 12:08:59.600495 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7246"] Dec 04 12:08:59 crc kubenswrapper[4979]: I1204 12:08:59.600714 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7246" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="registry-server" containerID="cri-o://4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9" gracePeriod=2 Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.018666 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.090719 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-catalog-content\") pod \"221352a3-a603-4a91-bba1-d7556bb94d16\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.090807 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkn8r\" (UniqueName: \"kubernetes.io/projected/221352a3-a603-4a91-bba1-d7556bb94d16-kube-api-access-mkn8r\") pod \"221352a3-a603-4a91-bba1-d7556bb94d16\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.090854 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-utilities\") pod \"221352a3-a603-4a91-bba1-d7556bb94d16\" (UID: \"221352a3-a603-4a91-bba1-d7556bb94d16\") " Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.092109 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-utilities" (OuterVolumeSpecName: "utilities") pod "221352a3-a603-4a91-bba1-d7556bb94d16" (UID: "221352a3-a603-4a91-bba1-d7556bb94d16"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.092489 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.096696 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/221352a3-a603-4a91-bba1-d7556bb94d16-kube-api-access-mkn8r" (OuterVolumeSpecName: "kube-api-access-mkn8r") pod "221352a3-a603-4a91-bba1-d7556bb94d16" (UID: "221352a3-a603-4a91-bba1-d7556bb94d16"). InnerVolumeSpecName "kube-api-access-mkn8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.148819 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "221352a3-a603-4a91-bba1-d7556bb94d16" (UID: "221352a3-a603-4a91-bba1-d7556bb94d16"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.194281 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/221352a3-a603-4a91-bba1-d7556bb94d16-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.194358 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkn8r\" (UniqueName: \"kubernetes.io/projected/221352a3-a603-4a91-bba1-d7556bb94d16-kube-api-access-mkn8r\") on node \"crc\" DevicePath \"\"" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.318586 4979 generic.go:334] "Generic (PLEG): container finished" podID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerID="dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430" exitCode=0 Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.318692 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2g6ml" event={"ID":"8aea49a0-460d-4dc2-8132-e13a34d4cced","Type":"ContainerDied","Data":"dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430"} Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.329173 4979 generic.go:334] "Generic (PLEG): container finished" podID="221352a3-a603-4a91-bba1-d7556bb94d16" containerID="4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9" exitCode=0 Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.329225 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7246" event={"ID":"221352a3-a603-4a91-bba1-d7556bb94d16","Type":"ContainerDied","Data":"4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9"} Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.329240 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7246" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.329268 4979 scope.go:117] "RemoveContainer" containerID="4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.329255 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7246" event={"ID":"221352a3-a603-4a91-bba1-d7556bb94d16","Type":"ContainerDied","Data":"9aa5ae4955a5cc05d17e26206953e921d3f9cdd1056e02ee26289e3612e5d653"} Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.360923 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7246"] Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.364788 4979 scope.go:117] "RemoveContainer" containerID="094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.367003 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7246"] Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.390257 4979 scope.go:117] "RemoveContainer" containerID="a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.409028 4979 scope.go:117] "RemoveContainer" containerID="4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9" Dec 04 12:09:00 crc kubenswrapper[4979]: E1204 12:09:00.409471 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9\": container with ID starting with 4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9 not found: ID does not exist" containerID="4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.409517 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9"} err="failed to get container status \"4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9\": rpc error: code = NotFound desc = could not find container \"4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9\": container with ID starting with 4acf1fcd6c6b3f4ae6e81144bf985848c295d71676930f0cdc89eaca907ad2a9 not found: ID does not exist" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.409548 4979 scope.go:117] "RemoveContainer" containerID="094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963" Dec 04 12:09:00 crc kubenswrapper[4979]: E1204 12:09:00.409924 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963\": container with ID starting with 094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963 not found: ID does not exist" containerID="094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.409951 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963"} err="failed to get container status \"094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963\": rpc error: code = NotFound desc = could not find container \"094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963\": container with ID starting with 094228994e5285ffc1ffc58a6ddf94c908fcdbdef8fe7dd6b25dc69d302b2963 not found: ID does not exist" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.409965 4979 scope.go:117] "RemoveContainer" containerID="a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65" Dec 04 12:09:00 crc kubenswrapper[4979]: E1204 12:09:00.410384 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65\": container with ID starting with a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65 not found: ID does not exist" containerID="a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65" Dec 04 12:09:00 crc kubenswrapper[4979]: I1204 12:09:00.410433 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65"} err="failed to get container status \"a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65\": rpc error: code = NotFound desc = could not find container \"a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65\": container with ID starting with a638f7b01c38d7607e91572d54494134ee35880ef52fce89555c625d303f5f65 not found: ID does not exist" Dec 04 12:09:01 crc kubenswrapper[4979]: I1204 12:09:01.342066 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2g6ml" event={"ID":"8aea49a0-460d-4dc2-8132-e13a34d4cced","Type":"ContainerStarted","Data":"c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f"} Dec 04 12:09:01 crc kubenswrapper[4979]: I1204 12:09:01.360497 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2g6ml" podStartSLOduration=3.905189764 podStartE2EDuration="5.360240123s" podCreationTimestamp="2025-12-04 12:08:56 +0000 UTC" firstStartedPulling="2025-12-04 12:08:59.305440322 +0000 UTC m=+1563.579736126" lastFinishedPulling="2025-12-04 12:09:00.760490641 +0000 UTC m=+1565.034786485" observedRunningTime="2025-12-04 12:09:01.358021864 +0000 UTC m=+1565.632317708" watchObservedRunningTime="2025-12-04 12:09:01.360240123 +0000 UTC m=+1565.634535927" Dec 04 12:09:02 crc kubenswrapper[4979]: I1204 12:09:02.047136 4979 scope.go:117] "RemoveContainer" containerID="fe043f249efb9dfd8bb0e3a20f1360262cf5b2c026e1870681a1fabdae2c45d6" Dec 04 12:09:02 crc kubenswrapper[4979]: I1204 12:09:02.070624 4979 scope.go:117] "RemoveContainer" containerID="5bfcf4a2c82014915164b9d068e35f70fe157a23fbb37df596106cebd300d320" Dec 04 12:09:02 crc kubenswrapper[4979]: I1204 12:09:02.104552 4979 scope.go:117] "RemoveContainer" containerID="7c15680c11567bfbc6297277c433787fb22a46e130fb931cdb93732c47665bb6" Dec 04 12:09:02 crc kubenswrapper[4979]: I1204 12:09:02.213026 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" path="/var/lib/kubelet/pods/221352a3-a603-4a91-bba1-d7556bb94d16/volumes" Dec 04 12:09:07 crc kubenswrapper[4979]: I1204 12:09:07.143373 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:09:07 crc kubenswrapper[4979]: I1204 12:09:07.143910 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:09:07 crc kubenswrapper[4979]: I1204 12:09:07.219637 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:09:07 crc kubenswrapper[4979]: I1204 12:09:07.466977 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:09:07 crc kubenswrapper[4979]: I1204 12:09:07.533116 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2g6ml"] Dec 04 12:09:08 crc kubenswrapper[4979]: I1204 12:09:08.199345 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:09:08 crc kubenswrapper[4979]: E1204 12:09:08.199650 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:09:09 crc kubenswrapper[4979]: I1204 12:09:09.418624 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2g6ml" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="registry-server" containerID="cri-o://c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f" gracePeriod=2 Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.377063 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.428530 4979 generic.go:334] "Generic (PLEG): container finished" podID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerID="c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f" exitCode=0 Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.428576 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2g6ml" event={"ID":"8aea49a0-460d-4dc2-8132-e13a34d4cced","Type":"ContainerDied","Data":"c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f"} Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.428582 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2g6ml" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.428606 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2g6ml" event={"ID":"8aea49a0-460d-4dc2-8132-e13a34d4cced","Type":"ContainerDied","Data":"9f85ba2be91e41e7277dbbcab1d302cdda4dae2ad92b1c95d611a5fbefaf2d03"} Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.428624 4979 scope.go:117] "RemoveContainer" containerID="c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.447390 4979 scope.go:117] "RemoveContainer" containerID="dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.452830 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvqvk\" (UniqueName: \"kubernetes.io/projected/8aea49a0-460d-4dc2-8132-e13a34d4cced-kube-api-access-bvqvk\") pod \"8aea49a0-460d-4dc2-8132-e13a34d4cced\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.452930 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-catalog-content\") pod \"8aea49a0-460d-4dc2-8132-e13a34d4cced\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.452988 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-utilities\") pod \"8aea49a0-460d-4dc2-8132-e13a34d4cced\" (UID: \"8aea49a0-460d-4dc2-8132-e13a34d4cced\") " Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.454185 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-utilities" (OuterVolumeSpecName: "utilities") pod "8aea49a0-460d-4dc2-8132-e13a34d4cced" (UID: "8aea49a0-460d-4dc2-8132-e13a34d4cced"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.458693 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aea49a0-460d-4dc2-8132-e13a34d4cced-kube-api-access-bvqvk" (OuterVolumeSpecName: "kube-api-access-bvqvk") pod "8aea49a0-460d-4dc2-8132-e13a34d4cced" (UID: "8aea49a0-460d-4dc2-8132-e13a34d4cced"). InnerVolumeSpecName "kube-api-access-bvqvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.465234 4979 scope.go:117] "RemoveContainer" containerID="bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.503134 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8aea49a0-460d-4dc2-8132-e13a34d4cced" (UID: "8aea49a0-460d-4dc2-8132-e13a34d4cced"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.505729 4979 scope.go:117] "RemoveContainer" containerID="c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f" Dec 04 12:09:10 crc kubenswrapper[4979]: E1204 12:09:10.506263 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f\": container with ID starting with c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f not found: ID does not exist" containerID="c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.506431 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f"} err="failed to get container status \"c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f\": rpc error: code = NotFound desc = could not find container \"c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f\": container with ID starting with c3c3c05d4cac27d6821af58f7d55b7020e7b450d7a0247059fdda9f4276e4e7f not found: ID does not exist" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.506521 4979 scope.go:117] "RemoveContainer" containerID="dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430" Dec 04 12:09:10 crc kubenswrapper[4979]: E1204 12:09:10.506886 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430\": container with ID starting with dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430 not found: ID does not exist" containerID="dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.506915 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430"} err="failed to get container status \"dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430\": rpc error: code = NotFound desc = could not find container \"dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430\": container with ID starting with dd0e39c04a346836aa77e88780e1646f68d43f598a2704edf1051d6fa45cf430 not found: ID does not exist" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.506929 4979 scope.go:117] "RemoveContainer" containerID="bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6" Dec 04 12:09:10 crc kubenswrapper[4979]: E1204 12:09:10.507206 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6\": container with ID starting with bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6 not found: ID does not exist" containerID="bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.507412 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6"} err="failed to get container status \"bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6\": rpc error: code = NotFound desc = could not find container \"bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6\": container with ID starting with bd2afc12f1437cef725ead78cd9ca37182ed069ce724c4da684210860958e1e6 not found: ID does not exist" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.553974 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.554007 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvqvk\" (UniqueName: \"kubernetes.io/projected/8aea49a0-460d-4dc2-8132-e13a34d4cced-kube-api-access-bvqvk\") on node \"crc\" DevicePath \"\"" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.554020 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aea49a0-460d-4dc2-8132-e13a34d4cced-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.761607 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2g6ml"] Dec 04 12:09:10 crc kubenswrapper[4979]: I1204 12:09:10.766615 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2g6ml"] Dec 04 12:09:12 crc kubenswrapper[4979]: I1204 12:09:12.210326 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" path="/var/lib/kubelet/pods/8aea49a0-460d-4dc2-8132-e13a34d4cced/volumes" Dec 04 12:09:20 crc kubenswrapper[4979]: I1204 12:09:20.199809 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:09:20 crc kubenswrapper[4979]: E1204 12:09:20.200745 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:09:33 crc kubenswrapper[4979]: I1204 12:09:33.199178 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:09:33 crc kubenswrapper[4979]: E1204 12:09:33.200278 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:09:46 crc kubenswrapper[4979]: I1204 12:09:46.209042 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:09:46 crc kubenswrapper[4979]: E1204 12:09:46.209581 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:09:58 crc kubenswrapper[4979]: I1204 12:09:58.200080 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:09:58 crc kubenswrapper[4979]: E1204 12:09:58.200902 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.602928 4979 scope.go:117] "RemoveContainer" containerID="01098447b1c91b182ef6963678cdbefbac832a1ebb6af0b9649ec51c2c66bb3b" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.633023 4979 scope.go:117] "RemoveContainer" containerID="a90821dd09be76982c32dd1c96a215f336b4994e444a92a4fa9851a9ea85a970" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.680506 4979 scope.go:117] "RemoveContainer" containerID="20abc3d60657fd8e86970f06db37aaca534a66acbeede0e238a1dc8d387e98fd" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.707875 4979 scope.go:117] "RemoveContainer" containerID="84ca39ca4cf3463b208bb9394e262a318f737de5230f3c751a488c39ef49a00f" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.741043 4979 scope.go:117] "RemoveContainer" containerID="14b0d7a646aefbe6dd898dc529926522899476170a2a82286a54b26b858ad281" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.777959 4979 scope.go:117] "RemoveContainer" containerID="abe4d0cf444f7acbf247979eb35483dd200bf8268c003927ff768849ca0addcc" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.808033 4979 scope.go:117] "RemoveContainer" containerID="992bc26c538b977fe854be958e44ced58bdcb0d9b0705720d5e5a3f11286c1a4" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.837714 4979 scope.go:117] "RemoveContainer" containerID="905bf4fc0f5406fa0d2d6f9b94de151c0ca8b3a78f05d8bb25b374f9b58efd38" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.861640 4979 scope.go:117] "RemoveContainer" containerID="b60e9bbefd212c20469dd92a8d6872954392efd8ce66ca080f0108e8f23c0598" Dec 04 12:10:02 crc kubenswrapper[4979]: I1204 12:10:02.883633 4979 scope.go:117] "RemoveContainer" containerID="3f0243e98bb73e51c6d7086f697067a94b8297badef0bd16f3868920634c8218" Dec 04 12:10:10 crc kubenswrapper[4979]: I1204 12:10:10.199705 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:10:10 crc kubenswrapper[4979]: E1204 12:10:10.200609 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:10:24 crc kubenswrapper[4979]: I1204 12:10:24.199740 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:10:24 crc kubenswrapper[4979]: E1204 12:10:24.201039 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:10:39 crc kubenswrapper[4979]: I1204 12:10:39.199787 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:10:39 crc kubenswrapper[4979]: E1204 12:10:39.200857 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:10:51 crc kubenswrapper[4979]: I1204 12:10:51.199727 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:10:51 crc kubenswrapper[4979]: E1204 12:10:51.200731 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.064275 4979 scope.go:117] "RemoveContainer" containerID="a386179a59dd07a6713af9202101515d8f0def69fccedf342f0246871f55cb4b" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.098336 4979 scope.go:117] "RemoveContainer" containerID="89858cce1ad717d7a0855a8184b9e88c16893052de81ca4a07efe6a69e691109" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.138031 4979 scope.go:117] "RemoveContainer" containerID="ddb76e21454d411642ff491c233f19bf9c4483d4330deff77afae441d539dc77" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.165484 4979 scope.go:117] "RemoveContainer" containerID="3854a6ae681a4089c468e8e386b10156644e06f17c8fc68fe7422dbcc59dbb35" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.195125 4979 scope.go:117] "RemoveContainer" containerID="dfee452098fffdbca65d5be95466eeb6448db8377ba02e117a1cd9cea7962b7f" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.231883 4979 scope.go:117] "RemoveContainer" containerID="1058b8753f8fe274cab6fb08732e2e77aada128e55d54b84e725482bf578e761" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.261731 4979 scope.go:117] "RemoveContainer" containerID="d36d0ca06131ebdb57c910e9f3a30e4390a3daaaeb7dc548462de91ece6eec7c" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.296107 4979 scope.go:117] "RemoveContainer" containerID="a1d5131cd2a4f7ced751fb83f28ba6274321f1df6ac8c22111baa1fa6c0440fc" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.329287 4979 scope.go:117] "RemoveContainer" containerID="e2e6214d14320ff207b7aebe8cd3428ad03cff707cf31a646cb28e7394285d50" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.351649 4979 scope.go:117] "RemoveContainer" containerID="07755584661f943a7836f5d09c354a242f8dc548a3c61640b8eac7753c254ff9" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.396628 4979 scope.go:117] "RemoveContainer" containerID="fc52fb357d41227cb0272a57c0f91beb6e46c9edf9d85b28eaeda62c1c8e1dfe" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.420560 4979 scope.go:117] "RemoveContainer" containerID="79e2c3ff30be15bd922b3998cf47243382b6d384d3d6b9b8c6698361d07eb8a1" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.446211 4979 scope.go:117] "RemoveContainer" containerID="3c64b869abf2343d534fe0e3db701fe908e10cca6fa7036f2bb28093bd4e5f7e" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.493328 4979 scope.go:117] "RemoveContainer" containerID="879ae343329fb352b347620cddbfda0a9c4a87f4868acb73890adbe05274d8d3" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.524856 4979 scope.go:117] "RemoveContainer" containerID="aed86b6e4e107649f8fb95f66dcae276fcc47522a30318a9927a27f1a5c5a593" Dec 04 12:11:03 crc kubenswrapper[4979]: I1204 12:11:03.588596 4979 scope.go:117] "RemoveContainer" containerID="bee5a7aae3b63933dd5a1e0a68fdee80d8892b52496944e2423977d5e0b997af" Dec 04 12:11:05 crc kubenswrapper[4979]: I1204 12:11:05.199526 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:11:05 crc kubenswrapper[4979]: E1204 12:11:05.200960 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:11:20 crc kubenswrapper[4979]: I1204 12:11:20.199413 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:11:20 crc kubenswrapper[4979]: E1204 12:11:20.200205 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:11:35 crc kubenswrapper[4979]: I1204 12:11:35.199349 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:11:35 crc kubenswrapper[4979]: E1204 12:11:35.200560 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:11:47 crc kubenswrapper[4979]: I1204 12:11:47.199136 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:11:47 crc kubenswrapper[4979]: E1204 12:11:47.199790 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:11:58 crc kubenswrapper[4979]: I1204 12:11:58.199072 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:11:58 crc kubenswrapper[4979]: E1204 12:11:58.199936 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.806984 4979 scope.go:117] "RemoveContainer" containerID="17d10021029714eaf81e19c13f5d0dfddee111a3c2ce627763576fdf61297d0c" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.833357 4979 scope.go:117] "RemoveContainer" containerID="0c6ee4d9251618d455bcbaeca27562a8a00ff00726d03c52d5bbc4be6e44f194" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.864386 4979 scope.go:117] "RemoveContainer" containerID="a07ddf9ea78efef5f6585f245d44e118f0a95206043e0fd406aef1c64f4cd3ca" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.896469 4979 scope.go:117] "RemoveContainer" containerID="d03c505af873328ff9c508812d00120785b8d279d1fa8b7860baaf6741c42a6c" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.929067 4979 scope.go:117] "RemoveContainer" containerID="3d54c5d9cb36dae2f6bcd838a3e0fddaf4c2d1f19e347ee987290f673c8efd37" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.956338 4979 scope.go:117] "RemoveContainer" containerID="aa4d07bf279adaae12d8aad041308e01478fd7338e4e1c8dfe3e93d2a095aa65" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.976195 4979 scope.go:117] "RemoveContainer" containerID="66265992f836c6a9f8e585ba56a557ed7f384463e217621e5223701c105f7441" Dec 04 12:12:03 crc kubenswrapper[4979]: I1204 12:12:03.996456 4979 scope.go:117] "RemoveContainer" containerID="1e209cfbc1ecb28128f57fb249b4b4a7d088a53a2531b190fcf5e7e253d5aed3" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.019330 4979 scope.go:117] "RemoveContainer" containerID="9acd8b6826bd47b62b1e082bc092cd3865472726e361e5dceb1e5f9d607411ba" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.042714 4979 scope.go:117] "RemoveContainer" containerID="453fe0a328e8492eb90c3463bc6db1ca1052fc0f0ce8b9544f2ae3008198ff55" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.065354 4979 scope.go:117] "RemoveContainer" containerID="0b18da01797ecd4c68f4b61b2e504256585316a4b410185ba40df8298d3b3107" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.087847 4979 scope.go:117] "RemoveContainer" containerID="bf93f510653a27db045fa5f965a3176f65591f23d896c18d1be176054af8b89f" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.104043 4979 scope.go:117] "RemoveContainer" containerID="02024a485457048e2f54ce565223a36a682c9da052f1a487dff4004a22a382ee" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.124899 4979 scope.go:117] "RemoveContainer" containerID="07d2a54c7f114041633c7190d149a251778f6bde1cdbf1ccdf26922e6d2aeed8" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.140282 4979 scope.go:117] "RemoveContainer" containerID="063e5850e136541b04a3fbcd523818e0db023f1015cc97b6258d7d2440f0fc84" Dec 04 12:12:04 crc kubenswrapper[4979]: I1204 12:12:04.158859 4979 scope.go:117] "RemoveContainer" containerID="7341c4702a0b8160bf9a2a2ae8f818d6b9089d0a90f6a4ed29d2510863a765d0" Dec 04 12:12:10 crc kubenswrapper[4979]: I1204 12:12:10.199973 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:12:10 crc kubenswrapper[4979]: E1204 12:12:10.200970 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:12:21 crc kubenswrapper[4979]: I1204 12:12:21.199647 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:12:21 crc kubenswrapper[4979]: E1204 12:12:21.200506 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:12:36 crc kubenswrapper[4979]: I1204 12:12:36.205687 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:12:36 crc kubenswrapper[4979]: E1204 12:12:36.206484 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:12:47 crc kubenswrapper[4979]: I1204 12:12:47.199370 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:12:47 crc kubenswrapper[4979]: E1204 12:12:47.199989 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:12:59 crc kubenswrapper[4979]: I1204 12:12:59.199428 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:12:59 crc kubenswrapper[4979]: E1204 12:12:59.200193 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:13:04 crc kubenswrapper[4979]: I1204 12:13:04.304209 4979 scope.go:117] "RemoveContainer" containerID="d6995eed3a9022adb589d900dcb50f240dcc437b7005c2eca54fd004c5fbffa7" Dec 04 12:13:04 crc kubenswrapper[4979]: I1204 12:13:04.356629 4979 scope.go:117] "RemoveContainer" containerID="716bf63a8969da92127307155cf7ce295dd76465e78ce82e446d08fcdb27f11a" Dec 04 12:13:04 crc kubenswrapper[4979]: I1204 12:13:04.390947 4979 scope.go:117] "RemoveContainer" containerID="6466c98b4d2b75be62b2edb44c9e2313bed08ae164cacc3957a45c111ebc7cc6" Dec 04 12:13:04 crc kubenswrapper[4979]: I1204 12:13:04.426128 4979 scope.go:117] "RemoveContainer" containerID="f17eef864c09a58ad72a34fd9c6c3191ef45028bc4ee2edec5f5297016f9e2a8" Dec 04 12:13:04 crc kubenswrapper[4979]: I1204 12:13:04.441114 4979 scope.go:117] "RemoveContainer" containerID="fa7c155782f73f5a00cf9804f2ca4eb191b625329d1e212029cc501c99a7deca" Dec 04 12:13:04 crc kubenswrapper[4979]: I1204 12:13:04.456233 4979 scope.go:117] "RemoveContainer" containerID="68bb06f1ab863da3c41eee87d1183cafa70bd684d96aec88da53971c3cdc2983" Dec 04 12:13:14 crc kubenswrapper[4979]: I1204 12:13:14.199341 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:13:14 crc kubenswrapper[4979]: E1204 12:13:14.200090 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:13:25 crc kubenswrapper[4979]: I1204 12:13:25.198733 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:13:25 crc kubenswrapper[4979]: E1204 12:13:25.199554 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:13:40 crc kubenswrapper[4979]: I1204 12:13:40.198898 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:13:40 crc kubenswrapper[4979]: I1204 12:13:40.828677 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"ddbc20e575f36c4d0f00640387a098ad7f525d1604cb7e2f7f21c76a51f78893"} Dec 04 12:14:04 crc kubenswrapper[4979]: I1204 12:14:04.540469 4979 scope.go:117] "RemoveContainer" containerID="4d858ae77dd1a31f97b848851105d4e11d7f14e77dcd0ed16e08b7acf7736bf5" Dec 04 12:14:04 crc kubenswrapper[4979]: I1204 12:14:04.587430 4979 scope.go:117] "RemoveContainer" containerID="676d3e7b21b50404e78713930c4c90e6f133e77d13cd221bda1604c38311a704" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140024 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm"] Dec 04 12:15:00 crc kubenswrapper[4979]: E1204 12:15:00.140862 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="extract-utilities" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140875 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="extract-utilities" Dec 04 12:15:00 crc kubenswrapper[4979]: E1204 12:15:00.140888 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="registry-server" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140895 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="registry-server" Dec 04 12:15:00 crc kubenswrapper[4979]: E1204 12:15:00.140908 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="extract-content" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140915 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="extract-content" Dec 04 12:15:00 crc kubenswrapper[4979]: E1204 12:15:00.140924 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="extract-utilities" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140931 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="extract-utilities" Dec 04 12:15:00 crc kubenswrapper[4979]: E1204 12:15:00.140946 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="extract-content" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140952 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="extract-content" Dec 04 12:15:00 crc kubenswrapper[4979]: E1204 12:15:00.140963 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="registry-server" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.140968 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="registry-server" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.141111 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="221352a3-a603-4a91-bba1-d7556bb94d16" containerName="registry-server" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.141144 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aea49a0-460d-4dc2-8132-e13a34d4cced" containerName="registry-server" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.141698 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.143893 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.145693 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.171724 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm"] Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.312789 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz6ds\" (UniqueName: \"kubernetes.io/projected/164112b0-ceef-4b99-8de0-fdd458ab6354-kube-api-access-sz6ds\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.313119 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164112b0-ceef-4b99-8de0-fdd458ab6354-secret-volume\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.313365 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164112b0-ceef-4b99-8de0-fdd458ab6354-config-volume\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.415354 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164112b0-ceef-4b99-8de0-fdd458ab6354-config-volume\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.415442 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz6ds\" (UniqueName: \"kubernetes.io/projected/164112b0-ceef-4b99-8de0-fdd458ab6354-kube-api-access-sz6ds\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.415535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164112b0-ceef-4b99-8de0-fdd458ab6354-secret-volume\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.416264 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164112b0-ceef-4b99-8de0-fdd458ab6354-config-volume\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.425330 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164112b0-ceef-4b99-8de0-fdd458ab6354-secret-volume\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.431619 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz6ds\" (UniqueName: \"kubernetes.io/projected/164112b0-ceef-4b99-8de0-fdd458ab6354-kube-api-access-sz6ds\") pod \"collect-profiles-29414175-w8gsm\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.468806 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:00 crc kubenswrapper[4979]: I1204 12:15:00.884703 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm"] Dec 04 12:15:01 crc kubenswrapper[4979]: I1204 12:15:01.414993 4979 generic.go:334] "Generic (PLEG): container finished" podID="164112b0-ceef-4b99-8de0-fdd458ab6354" containerID="594a1a3de87a48be0a9ccf3c5c5b9e4706868d1fc7c08f67a295e507a14d6367" exitCode=0 Dec 04 12:15:01 crc kubenswrapper[4979]: I1204 12:15:01.415048 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" event={"ID":"164112b0-ceef-4b99-8de0-fdd458ab6354","Type":"ContainerDied","Data":"594a1a3de87a48be0a9ccf3c5c5b9e4706868d1fc7c08f67a295e507a14d6367"} Dec 04 12:15:01 crc kubenswrapper[4979]: I1204 12:15:01.415081 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" event={"ID":"164112b0-ceef-4b99-8de0-fdd458ab6354","Type":"ContainerStarted","Data":"a7c8b76d2cece2a6004f5d022670e41bf1a9da9ab509bf22cc0318093d8c38f4"} Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.674206 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.847509 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164112b0-ceef-4b99-8de0-fdd458ab6354-secret-volume\") pod \"164112b0-ceef-4b99-8de0-fdd458ab6354\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.847604 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164112b0-ceef-4b99-8de0-fdd458ab6354-config-volume\") pod \"164112b0-ceef-4b99-8de0-fdd458ab6354\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.847715 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sz6ds\" (UniqueName: \"kubernetes.io/projected/164112b0-ceef-4b99-8de0-fdd458ab6354-kube-api-access-sz6ds\") pod \"164112b0-ceef-4b99-8de0-fdd458ab6354\" (UID: \"164112b0-ceef-4b99-8de0-fdd458ab6354\") " Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.848734 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/164112b0-ceef-4b99-8de0-fdd458ab6354-config-volume" (OuterVolumeSpecName: "config-volume") pod "164112b0-ceef-4b99-8de0-fdd458ab6354" (UID: "164112b0-ceef-4b99-8de0-fdd458ab6354"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.854871 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/164112b0-ceef-4b99-8de0-fdd458ab6354-kube-api-access-sz6ds" (OuterVolumeSpecName: "kube-api-access-sz6ds") pod "164112b0-ceef-4b99-8de0-fdd458ab6354" (UID: "164112b0-ceef-4b99-8de0-fdd458ab6354"). InnerVolumeSpecName "kube-api-access-sz6ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.855035 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/164112b0-ceef-4b99-8de0-fdd458ab6354-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "164112b0-ceef-4b99-8de0-fdd458ab6354" (UID: "164112b0-ceef-4b99-8de0-fdd458ab6354"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.950334 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/164112b0-ceef-4b99-8de0-fdd458ab6354-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.950429 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/164112b0-ceef-4b99-8de0-fdd458ab6354-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:15:02 crc kubenswrapper[4979]: I1204 12:15:02.950458 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sz6ds\" (UniqueName: \"kubernetes.io/projected/164112b0-ceef-4b99-8de0-fdd458ab6354-kube-api-access-sz6ds\") on node \"crc\" DevicePath \"\"" Dec 04 12:15:03 crc kubenswrapper[4979]: I1204 12:15:03.431643 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" event={"ID":"164112b0-ceef-4b99-8de0-fdd458ab6354","Type":"ContainerDied","Data":"a7c8b76d2cece2a6004f5d022670e41bf1a9da9ab509bf22cc0318093d8c38f4"} Dec 04 12:15:03 crc kubenswrapper[4979]: I1204 12:15:03.432028 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7c8b76d2cece2a6004f5d022670e41bf1a9da9ab509bf22cc0318093d8c38f4" Dec 04 12:15:03 crc kubenswrapper[4979]: I1204 12:15:03.431729 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm" Dec 04 12:15:04 crc kubenswrapper[4979]: I1204 12:15:04.647624 4979 scope.go:117] "RemoveContainer" containerID="dfb99c9b2fd24e29118b00f0d05832ad1a6f5468ea6b2c9b2e35da8ba48334d2" Dec 04 12:15:04 crc kubenswrapper[4979]: I1204 12:15:04.665713 4979 scope.go:117] "RemoveContainer" containerID="e0d2fdaa377a63293816ad71e539f322b7ffe64f4aa7fab9b51cdf23832d0f21" Dec 04 12:15:04 crc kubenswrapper[4979]: I1204 12:15:04.689747 4979 scope.go:117] "RemoveContainer" containerID="5c5f1bce73c540715d7e8d457680551f4966428323cb84c1dbe1d33aa8ee2de6" Dec 04 12:15:58 crc kubenswrapper[4979]: I1204 12:15:58.040561 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:15:58 crc kubenswrapper[4979]: I1204 12:15:58.041041 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:16:28 crc kubenswrapper[4979]: I1204 12:16:28.041177 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:16:28 crc kubenswrapper[4979]: I1204 12:16:28.041943 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.041270 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.042058 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.042119 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.042931 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ddbc20e575f36c4d0f00640387a098ad7f525d1604cb7e2f7f21c76a51f78893"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.043013 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://ddbc20e575f36c4d0f00640387a098ad7f525d1604cb7e2f7f21c76a51f78893" gracePeriod=600 Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.388723 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="ddbc20e575f36c4d0f00640387a098ad7f525d1604cb7e2f7f21c76a51f78893" exitCode=0 Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.388806 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"ddbc20e575f36c4d0f00640387a098ad7f525d1604cb7e2f7f21c76a51f78893"} Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.389274 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31"} Dec 04 12:16:58 crc kubenswrapper[4979]: I1204 12:16:58.389343 4979 scope.go:117] "RemoveContainer" containerID="350d1f7ed1952e77e30b9b7b9eef5573e8e197cd319ad66c05604ad0be5c1bbf" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.295420 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qr9lm"] Dec 04 12:17:38 crc kubenswrapper[4979]: E1204 12:17:38.296544 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="164112b0-ceef-4b99-8de0-fdd458ab6354" containerName="collect-profiles" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.296565 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="164112b0-ceef-4b99-8de0-fdd458ab6354" containerName="collect-profiles" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.296732 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="164112b0-ceef-4b99-8de0-fdd458ab6354" containerName="collect-profiles" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.300870 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.321468 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qr9lm"] Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.427570 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-utilities\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.427631 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7lfq\" (UniqueName: \"kubernetes.io/projected/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-kube-api-access-q7lfq\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.427741 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-catalog-content\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.528696 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-utilities\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.528766 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7lfq\" (UniqueName: \"kubernetes.io/projected/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-kube-api-access-q7lfq\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.528848 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-catalog-content\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.529162 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-utilities\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.529368 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-catalog-content\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.551175 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7lfq\" (UniqueName: \"kubernetes.io/projected/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-kube-api-access-q7lfq\") pod \"redhat-operators-qr9lm\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.630293 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:38 crc kubenswrapper[4979]: I1204 12:17:38.871073 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qr9lm"] Dec 04 12:17:39 crc kubenswrapper[4979]: I1204 12:17:39.741138 4979 generic.go:334] "Generic (PLEG): container finished" podID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerID="9ca7d91a85b6cc62937d1ad217b45f6518a2d6e46fadd7c7434a92fd12403b5d" exitCode=0 Dec 04 12:17:39 crc kubenswrapper[4979]: I1204 12:17:39.741200 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerDied","Data":"9ca7d91a85b6cc62937d1ad217b45f6518a2d6e46fadd7c7434a92fd12403b5d"} Dec 04 12:17:39 crc kubenswrapper[4979]: I1204 12:17:39.741239 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerStarted","Data":"36b6f59826717a7e9407a60e103b864edf8f110c75f36e6a44830566dc223ff5"} Dec 04 12:17:39 crc kubenswrapper[4979]: I1204 12:17:39.743706 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:17:40 crc kubenswrapper[4979]: I1204 12:17:40.748195 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerStarted","Data":"d0ef82977214dddeddffb602e232d044358a4410c2d5c32b4f2fd1bec16306c7"} Dec 04 12:17:41 crc kubenswrapper[4979]: I1204 12:17:41.760833 4979 generic.go:334] "Generic (PLEG): container finished" podID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerID="d0ef82977214dddeddffb602e232d044358a4410c2d5c32b4f2fd1bec16306c7" exitCode=0 Dec 04 12:17:41 crc kubenswrapper[4979]: I1204 12:17:41.760885 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerDied","Data":"d0ef82977214dddeddffb602e232d044358a4410c2d5c32b4f2fd1bec16306c7"} Dec 04 12:17:42 crc kubenswrapper[4979]: I1204 12:17:42.770243 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerStarted","Data":"1351df846bf0d97defb209b685adf15674358fa880668081af8cff0ab120207b"} Dec 04 12:17:42 crc kubenswrapper[4979]: I1204 12:17:42.790096 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qr9lm" podStartSLOduration=2.205463642 podStartE2EDuration="4.790076391s" podCreationTimestamp="2025-12-04 12:17:38 +0000 UTC" firstStartedPulling="2025-12-04 12:17:39.743271605 +0000 UTC m=+2084.017567439" lastFinishedPulling="2025-12-04 12:17:42.327884384 +0000 UTC m=+2086.602180188" observedRunningTime="2025-12-04 12:17:42.78527224 +0000 UTC m=+2087.059568064" watchObservedRunningTime="2025-12-04 12:17:42.790076391 +0000 UTC m=+2087.064372205" Dec 04 12:17:48 crc kubenswrapper[4979]: I1204 12:17:48.631351 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:48 crc kubenswrapper[4979]: I1204 12:17:48.631901 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:48 crc kubenswrapper[4979]: I1204 12:17:48.676560 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:48 crc kubenswrapper[4979]: I1204 12:17:48.868706 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:48 crc kubenswrapper[4979]: I1204 12:17:48.910457 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qr9lm"] Dec 04 12:17:50 crc kubenswrapper[4979]: I1204 12:17:50.829682 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qr9lm" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="registry-server" containerID="cri-o://1351df846bf0d97defb209b685adf15674358fa880668081af8cff0ab120207b" gracePeriod=2 Dec 04 12:17:53 crc kubenswrapper[4979]: I1204 12:17:53.852763 4979 generic.go:334] "Generic (PLEG): container finished" podID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerID="1351df846bf0d97defb209b685adf15674358fa880668081af8cff0ab120207b" exitCode=0 Dec 04 12:17:53 crc kubenswrapper[4979]: I1204 12:17:53.852860 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerDied","Data":"1351df846bf0d97defb209b685adf15674358fa880668081af8cff0ab120207b"} Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.620036 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.661687 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-catalog-content\") pod \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.661759 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-utilities\") pod \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.661847 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7lfq\" (UniqueName: \"kubernetes.io/projected/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-kube-api-access-q7lfq\") pod \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\" (UID: \"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13\") " Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.662751 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-utilities" (OuterVolumeSpecName: "utilities") pod "fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" (UID: "fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.667033 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-kube-api-access-q7lfq" (OuterVolumeSpecName: "kube-api-access-q7lfq") pod "fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" (UID: "fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13"). InnerVolumeSpecName "kube-api-access-q7lfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.764109 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7lfq\" (UniqueName: \"kubernetes.io/projected/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-kube-api-access-q7lfq\") on node \"crc\" DevicePath \"\"" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.764148 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.780801 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" (UID: "fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.862100 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qr9lm" event={"ID":"fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13","Type":"ContainerDied","Data":"36b6f59826717a7e9407a60e103b864edf8f110c75f36e6a44830566dc223ff5"} Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.862141 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qr9lm" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.862164 4979 scope.go:117] "RemoveContainer" containerID="1351df846bf0d97defb209b685adf15674358fa880668081af8cff0ab120207b" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.865110 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.885033 4979 scope.go:117] "RemoveContainer" containerID="d0ef82977214dddeddffb602e232d044358a4410c2d5c32b4f2fd1bec16306c7" Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.895041 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qr9lm"] Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.901764 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qr9lm"] Dec 04 12:17:54 crc kubenswrapper[4979]: I1204 12:17:54.927843 4979 scope.go:117] "RemoveContainer" containerID="9ca7d91a85b6cc62937d1ad217b45f6518a2d6e46fadd7c7434a92fd12403b5d" Dec 04 12:17:56 crc kubenswrapper[4979]: I1204 12:17:56.209903 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" path="/var/lib/kubelet/pods/fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13/volumes" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.033910 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5sz8h"] Dec 04 12:18:18 crc kubenswrapper[4979]: E1204 12:18:18.034972 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="registry-server" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.034992 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="registry-server" Dec 04 12:18:18 crc kubenswrapper[4979]: E1204 12:18:18.035018 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="extract-utilities" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.035032 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="extract-utilities" Dec 04 12:18:18 crc kubenswrapper[4979]: E1204 12:18:18.035057 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="extract-content" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.035070 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="extract-content" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.035388 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdc7a5d1-4c3b-4cbd-8a5d-accabc37ec13" containerName="registry-server" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.037112 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.053350 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sz8h"] Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.097942 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85vpf\" (UniqueName: \"kubernetes.io/projected/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-kube-api-access-85vpf\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.098115 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-utilities\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.098225 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-catalog-content\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.199623 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85vpf\" (UniqueName: \"kubernetes.io/projected/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-kube-api-access-85vpf\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.199699 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-utilities\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.199750 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-catalog-content\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.200232 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-catalog-content\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.200593 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-utilities\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.221849 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85vpf\" (UniqueName: \"kubernetes.io/projected/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-kube-api-access-85vpf\") pod \"redhat-marketplace-5sz8h\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.363674 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:18 crc kubenswrapper[4979]: I1204 12:18:18.806659 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sz8h"] Dec 04 12:18:19 crc kubenswrapper[4979]: I1204 12:18:19.054546 4979 generic.go:334] "Generic (PLEG): container finished" podID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerID="58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9" exitCode=0 Dec 04 12:18:19 crc kubenswrapper[4979]: I1204 12:18:19.054610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sz8h" event={"ID":"c3e54d9b-8a56-4641-8ab4-24e0400ea19b","Type":"ContainerDied","Data":"58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9"} Dec 04 12:18:19 crc kubenswrapper[4979]: I1204 12:18:19.054649 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sz8h" event={"ID":"c3e54d9b-8a56-4641-8ab4-24e0400ea19b","Type":"ContainerStarted","Data":"9e3ce1607dafd0b23cf420d896b6d18b6002676e93e8d19a37727a65b185b846"} Dec 04 12:18:21 crc kubenswrapper[4979]: I1204 12:18:21.071627 4979 generic.go:334] "Generic (PLEG): container finished" podID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerID="e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60" exitCode=0 Dec 04 12:18:21 crc kubenswrapper[4979]: I1204 12:18:21.071710 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sz8h" event={"ID":"c3e54d9b-8a56-4641-8ab4-24e0400ea19b","Type":"ContainerDied","Data":"e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60"} Dec 04 12:18:22 crc kubenswrapper[4979]: I1204 12:18:22.081763 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sz8h" event={"ID":"c3e54d9b-8a56-4641-8ab4-24e0400ea19b","Type":"ContainerStarted","Data":"4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5"} Dec 04 12:18:22 crc kubenswrapper[4979]: I1204 12:18:22.107688 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5sz8h" podStartSLOduration=1.6543581 podStartE2EDuration="4.107666747s" podCreationTimestamp="2025-12-04 12:18:18 +0000 UTC" firstStartedPulling="2025-12-04 12:18:19.055939527 +0000 UTC m=+2123.330235331" lastFinishedPulling="2025-12-04 12:18:21.509248184 +0000 UTC m=+2125.783543978" observedRunningTime="2025-12-04 12:18:22.101328735 +0000 UTC m=+2126.375624539" watchObservedRunningTime="2025-12-04 12:18:22.107666747 +0000 UTC m=+2126.381962551" Dec 04 12:18:28 crc kubenswrapper[4979]: I1204 12:18:28.364688 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:28 crc kubenswrapper[4979]: I1204 12:18:28.365269 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:28 crc kubenswrapper[4979]: I1204 12:18:28.416575 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:29 crc kubenswrapper[4979]: I1204 12:18:29.181289 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:29 crc kubenswrapper[4979]: I1204 12:18:29.226122 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sz8h"] Dec 04 12:18:31 crc kubenswrapper[4979]: I1204 12:18:31.141373 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5sz8h" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="registry-server" containerID="cri-o://4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5" gracePeriod=2 Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.100784 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.149869 4979 generic.go:334] "Generic (PLEG): container finished" podID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerID="4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5" exitCode=0 Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.149916 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sz8h" event={"ID":"c3e54d9b-8a56-4641-8ab4-24e0400ea19b","Type":"ContainerDied","Data":"4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5"} Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.149944 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5sz8h" event={"ID":"c3e54d9b-8a56-4641-8ab4-24e0400ea19b","Type":"ContainerDied","Data":"9e3ce1607dafd0b23cf420d896b6d18b6002676e93e8d19a37727a65b185b846"} Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.149963 4979 scope.go:117] "RemoveContainer" containerID="4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.150100 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5sz8h" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.168112 4979 scope.go:117] "RemoveContainer" containerID="e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.185798 4979 scope.go:117] "RemoveContainer" containerID="58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.199088 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85vpf\" (UniqueName: \"kubernetes.io/projected/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-kube-api-access-85vpf\") pod \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.199251 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-catalog-content\") pod \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.199361 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-utilities\") pod \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\" (UID: \"c3e54d9b-8a56-4641-8ab4-24e0400ea19b\") " Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.200728 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-utilities" (OuterVolumeSpecName: "utilities") pod "c3e54d9b-8a56-4641-8ab4-24e0400ea19b" (UID: "c3e54d9b-8a56-4641-8ab4-24e0400ea19b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.208809 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-kube-api-access-85vpf" (OuterVolumeSpecName: "kube-api-access-85vpf") pod "c3e54d9b-8a56-4641-8ab4-24e0400ea19b" (UID: "c3e54d9b-8a56-4641-8ab4-24e0400ea19b"). InnerVolumeSpecName "kube-api-access-85vpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.214218 4979 scope.go:117] "RemoveContainer" containerID="4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5" Dec 04 12:18:32 crc kubenswrapper[4979]: E1204 12:18:32.215577 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5\": container with ID starting with 4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5 not found: ID does not exist" containerID="4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.215619 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5"} err="failed to get container status \"4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5\": rpc error: code = NotFound desc = could not find container \"4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5\": container with ID starting with 4626f404745640754b3e1d6c45b75419ad5814e4425f19a4d6d74f48e5e528d5 not found: ID does not exist" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.215645 4979 scope.go:117] "RemoveContainer" containerID="e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60" Dec 04 12:18:32 crc kubenswrapper[4979]: E1204 12:18:32.215970 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60\": container with ID starting with e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60 not found: ID does not exist" containerID="e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.216024 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60"} err="failed to get container status \"e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60\": rpc error: code = NotFound desc = could not find container \"e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60\": container with ID starting with e0d3e3840bf6eff997d2ac0273b0f193c2f1fa3ceee2420bbc4404006dee0e60 not found: ID does not exist" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.216059 4979 scope.go:117] "RemoveContainer" containerID="58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9" Dec 04 12:18:32 crc kubenswrapper[4979]: E1204 12:18:32.217497 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9\": container with ID starting with 58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9 not found: ID does not exist" containerID="58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.217542 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9"} err="failed to get container status \"58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9\": rpc error: code = NotFound desc = could not find container \"58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9\": container with ID starting with 58147b14070c8554fd908a6efe43b591b59e1160b20533128181ef73a174ddb9 not found: ID does not exist" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.225122 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c3e54d9b-8a56-4641-8ab4-24e0400ea19b" (UID: "c3e54d9b-8a56-4641-8ab4-24e0400ea19b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.300605 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.300653 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85vpf\" (UniqueName: \"kubernetes.io/projected/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-kube-api-access-85vpf\") on node \"crc\" DevicePath \"\"" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.300665 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c3e54d9b-8a56-4641-8ab4-24e0400ea19b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.483328 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sz8h"] Dec 04 12:18:32 crc kubenswrapper[4979]: I1204 12:18:32.488849 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5sz8h"] Dec 04 12:18:34 crc kubenswrapper[4979]: I1204 12:18:34.209386 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" path="/var/lib/kubelet/pods/c3e54d9b-8a56-4641-8ab4-24e0400ea19b/volumes" Dec 04 12:18:58 crc kubenswrapper[4979]: I1204 12:18:58.040945 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:18:58 crc kubenswrapper[4979]: I1204 12:18:58.041619 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:19:28 crc kubenswrapper[4979]: I1204 12:19:28.040927 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:19:28 crc kubenswrapper[4979]: I1204 12:19:28.042007 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:19:58 crc kubenswrapper[4979]: I1204 12:19:58.040754 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:19:58 crc kubenswrapper[4979]: I1204 12:19:58.041423 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:19:58 crc kubenswrapper[4979]: I1204 12:19:58.041484 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:19:58 crc kubenswrapper[4979]: I1204 12:19:58.042160 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:19:58 crc kubenswrapper[4979]: I1204 12:19:58.042221 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" gracePeriod=600 Dec 04 12:19:58 crc kubenswrapper[4979]: E1204 12:19:58.174974 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:19:59 crc kubenswrapper[4979]: I1204 12:19:59.077673 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" exitCode=0 Dec 04 12:19:59 crc kubenswrapper[4979]: I1204 12:19:59.077749 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31"} Dec 04 12:19:59 crc kubenswrapper[4979]: I1204 12:19:59.077796 4979 scope.go:117] "RemoveContainer" containerID="ddbc20e575f36c4d0f00640387a098ad7f525d1604cb7e2f7f21c76a51f78893" Dec 04 12:19:59 crc kubenswrapper[4979]: I1204 12:19:59.078565 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:19:59 crc kubenswrapper[4979]: E1204 12:19:59.079039 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:20:12 crc kubenswrapper[4979]: I1204 12:20:12.199787 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:20:12 crc kubenswrapper[4979]: E1204 12:20:12.200747 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:20:24 crc kubenswrapper[4979]: I1204 12:20:24.199368 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:20:24 crc kubenswrapper[4979]: E1204 12:20:24.200339 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:20:38 crc kubenswrapper[4979]: I1204 12:20:38.198962 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:20:38 crc kubenswrapper[4979]: E1204 12:20:38.199804 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:20:51 crc kubenswrapper[4979]: I1204 12:20:51.199099 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:20:51 crc kubenswrapper[4979]: E1204 12:20:51.200269 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:21:03 crc kubenswrapper[4979]: I1204 12:21:03.199633 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:21:03 crc kubenswrapper[4979]: E1204 12:21:03.200984 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:21:15 crc kubenswrapper[4979]: I1204 12:21:15.198886 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:21:15 crc kubenswrapper[4979]: E1204 12:21:15.199585 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:21:27 crc kubenswrapper[4979]: I1204 12:21:27.198701 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:21:27 crc kubenswrapper[4979]: E1204 12:21:27.199758 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:21:38 crc kubenswrapper[4979]: I1204 12:21:38.198996 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:21:38 crc kubenswrapper[4979]: E1204 12:21:38.199758 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:21:51 crc kubenswrapper[4979]: I1204 12:21:51.199193 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:21:51 crc kubenswrapper[4979]: E1204 12:21:51.200364 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.634941 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rv9jd"] Dec 04 12:21:53 crc kubenswrapper[4979]: E1204 12:21:53.639896 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="registry-server" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.639937 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="registry-server" Dec 04 12:21:53 crc kubenswrapper[4979]: E1204 12:21:53.639955 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="extract-content" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.639963 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="extract-content" Dec 04 12:21:53 crc kubenswrapper[4979]: E1204 12:21:53.640010 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="extract-utilities" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.640019 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="extract-utilities" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.640265 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3e54d9b-8a56-4641-8ab4-24e0400ea19b" containerName="registry-server" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.641642 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.651132 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rv9jd"] Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.767147 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-catalog-content\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.767262 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-utilities\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.767379 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g65wt\" (UniqueName: \"kubernetes.io/projected/44393a30-a46e-4ba0-ae46-e08499ddf29f-kube-api-access-g65wt\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.869015 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-catalog-content\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.869064 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-utilities\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.869095 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g65wt\" (UniqueName: \"kubernetes.io/projected/44393a30-a46e-4ba0-ae46-e08499ddf29f-kube-api-access-g65wt\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.869649 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-catalog-content\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.869731 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-utilities\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.898731 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g65wt\" (UniqueName: \"kubernetes.io/projected/44393a30-a46e-4ba0-ae46-e08499ddf29f-kube-api-access-g65wt\") pod \"certified-operators-rv9jd\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:53 crc kubenswrapper[4979]: I1204 12:21:53.961931 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:21:54 crc kubenswrapper[4979]: I1204 12:21:54.252005 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rv9jd"] Dec 04 12:21:54 crc kubenswrapper[4979]: I1204 12:21:54.999114 4979 generic.go:334] "Generic (PLEG): container finished" podID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerID="4b89f36ebdee07c018abbf3176ae4243abaf3d483233cb79eb10806dda10404b" exitCode=0 Dec 04 12:21:54 crc kubenswrapper[4979]: I1204 12:21:54.999159 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerDied","Data":"4b89f36ebdee07c018abbf3176ae4243abaf3d483233cb79eb10806dda10404b"} Dec 04 12:21:54 crc kubenswrapper[4979]: I1204 12:21:54.999468 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerStarted","Data":"b50698124fc1d696c084baee81f667e4673b94af0d1ba6e6e53973e3ce3382cc"} Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.413039 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p4j49"] Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.414381 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.431449 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4j49"] Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.591929 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j52c2\" (UniqueName: \"kubernetes.io/projected/618e561d-dcea-4355-931a-c80392f77d66-kube-api-access-j52c2\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.592007 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-utilities\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.592181 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-catalog-content\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.693147 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-utilities\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.693562 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-catalog-content\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.693613 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j52c2\" (UniqueName: \"kubernetes.io/projected/618e561d-dcea-4355-931a-c80392f77d66-kube-api-access-j52c2\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.693832 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-utilities\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.694018 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-catalog-content\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.713000 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j52c2\" (UniqueName: \"kubernetes.io/projected/618e561d-dcea-4355-931a-c80392f77d66-kube-api-access-j52c2\") pod \"community-operators-p4j49\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:55 crc kubenswrapper[4979]: I1204 12:21:55.733742 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:21:56 crc kubenswrapper[4979]: I1204 12:21:56.011519 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerStarted","Data":"710062232b1ab1039f70a0ec78f0e357575d0340d09c4f81f43e4705c2122dd1"} Dec 04 12:21:56 crc kubenswrapper[4979]: I1204 12:21:56.209495 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p4j49"] Dec 04 12:21:56 crc kubenswrapper[4979]: W1204 12:21:56.239363 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod618e561d_dcea_4355_931a_c80392f77d66.slice/crio-ccb7bb3db592371b1f6a2e41b06a2e15471bcda78c1a0ce1d1f5dd92e7720552 WatchSource:0}: Error finding container ccb7bb3db592371b1f6a2e41b06a2e15471bcda78c1a0ce1d1f5dd92e7720552: Status 404 returned error can't find the container with id ccb7bb3db592371b1f6a2e41b06a2e15471bcda78c1a0ce1d1f5dd92e7720552 Dec 04 12:21:57 crc kubenswrapper[4979]: I1204 12:21:57.026457 4979 generic.go:334] "Generic (PLEG): container finished" podID="618e561d-dcea-4355-931a-c80392f77d66" containerID="bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98" exitCode=0 Dec 04 12:21:57 crc kubenswrapper[4979]: I1204 12:21:57.026527 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerDied","Data":"bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98"} Dec 04 12:21:57 crc kubenswrapper[4979]: I1204 12:21:57.026896 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerStarted","Data":"ccb7bb3db592371b1f6a2e41b06a2e15471bcda78c1a0ce1d1f5dd92e7720552"} Dec 04 12:21:57 crc kubenswrapper[4979]: I1204 12:21:57.029687 4979 generic.go:334] "Generic (PLEG): container finished" podID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerID="710062232b1ab1039f70a0ec78f0e357575d0340d09c4f81f43e4705c2122dd1" exitCode=0 Dec 04 12:21:57 crc kubenswrapper[4979]: I1204 12:21:57.029746 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerDied","Data":"710062232b1ab1039f70a0ec78f0e357575d0340d09c4f81f43e4705c2122dd1"} Dec 04 12:21:58 crc kubenswrapper[4979]: I1204 12:21:58.040235 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerStarted","Data":"87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac"} Dec 04 12:21:58 crc kubenswrapper[4979]: I1204 12:21:58.043712 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerStarted","Data":"cf225ec7a80620e2281d80aaf696da33977aeab62e585c03c93e163ebfb43f5c"} Dec 04 12:21:58 crc kubenswrapper[4979]: I1204 12:21:58.096498 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rv9jd" podStartSLOduration=2.6626138409999998 podStartE2EDuration="5.096473183s" podCreationTimestamp="2025-12-04 12:21:53 +0000 UTC" firstStartedPulling="2025-12-04 12:21:55.000721899 +0000 UTC m=+2339.275017713" lastFinishedPulling="2025-12-04 12:21:57.434581251 +0000 UTC m=+2341.708877055" observedRunningTime="2025-12-04 12:21:58.093146422 +0000 UTC m=+2342.367442226" watchObservedRunningTime="2025-12-04 12:21:58.096473183 +0000 UTC m=+2342.370768987" Dec 04 12:21:59 crc kubenswrapper[4979]: I1204 12:21:59.052701 4979 generic.go:334] "Generic (PLEG): container finished" podID="618e561d-dcea-4355-931a-c80392f77d66" containerID="87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac" exitCode=0 Dec 04 12:21:59 crc kubenswrapper[4979]: I1204 12:21:59.052858 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerDied","Data":"87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac"} Dec 04 12:22:00 crc kubenswrapper[4979]: I1204 12:22:00.064385 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerStarted","Data":"cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f"} Dec 04 12:22:00 crc kubenswrapper[4979]: I1204 12:22:00.090319 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p4j49" podStartSLOduration=2.657927548 podStartE2EDuration="5.090280119s" podCreationTimestamp="2025-12-04 12:21:55 +0000 UTC" firstStartedPulling="2025-12-04 12:21:57.029058913 +0000 UTC m=+2341.303354727" lastFinishedPulling="2025-12-04 12:21:59.461411494 +0000 UTC m=+2343.735707298" observedRunningTime="2025-12-04 12:22:00.084030589 +0000 UTC m=+2344.358326453" watchObservedRunningTime="2025-12-04 12:22:00.090280119 +0000 UTC m=+2344.364575943" Dec 04 12:22:03 crc kubenswrapper[4979]: I1204 12:22:03.962817 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:22:03 crc kubenswrapper[4979]: I1204 12:22:03.963235 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:22:04 crc kubenswrapper[4979]: I1204 12:22:04.008674 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:22:04 crc kubenswrapper[4979]: I1204 12:22:04.155497 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:22:05 crc kubenswrapper[4979]: I1204 12:22:05.009847 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rv9jd"] Dec 04 12:22:05 crc kubenswrapper[4979]: I1204 12:22:05.199086 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:22:05 crc kubenswrapper[4979]: E1204 12:22:05.199501 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:22:05 crc kubenswrapper[4979]: I1204 12:22:05.734630 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:22:05 crc kubenswrapper[4979]: I1204 12:22:05.734782 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:22:05 crc kubenswrapper[4979]: I1204 12:22:05.778563 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:22:06 crc kubenswrapper[4979]: I1204 12:22:06.116503 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rv9jd" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="registry-server" containerID="cri-o://cf225ec7a80620e2281d80aaf696da33977aeab62e585c03c93e163ebfb43f5c" gracePeriod=2 Dec 04 12:22:06 crc kubenswrapper[4979]: I1204 12:22:06.164290 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:22:07 crc kubenswrapper[4979]: I1204 12:22:07.409754 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4j49"] Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.128914 4979 generic.go:334] "Generic (PLEG): container finished" podID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerID="cf225ec7a80620e2281d80aaf696da33977aeab62e585c03c93e163ebfb43f5c" exitCode=0 Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.129078 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerDied","Data":"cf225ec7a80620e2281d80aaf696da33977aeab62e585c03c93e163ebfb43f5c"} Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.683986 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.774242 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g65wt\" (UniqueName: \"kubernetes.io/projected/44393a30-a46e-4ba0-ae46-e08499ddf29f-kube-api-access-g65wt\") pod \"44393a30-a46e-4ba0-ae46-e08499ddf29f\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.774352 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-catalog-content\") pod \"44393a30-a46e-4ba0-ae46-e08499ddf29f\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.774393 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-utilities\") pod \"44393a30-a46e-4ba0-ae46-e08499ddf29f\" (UID: \"44393a30-a46e-4ba0-ae46-e08499ddf29f\") " Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.775665 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-utilities" (OuterVolumeSpecName: "utilities") pod "44393a30-a46e-4ba0-ae46-e08499ddf29f" (UID: "44393a30-a46e-4ba0-ae46-e08499ddf29f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.779738 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44393a30-a46e-4ba0-ae46-e08499ddf29f-kube-api-access-g65wt" (OuterVolumeSpecName: "kube-api-access-g65wt") pod "44393a30-a46e-4ba0-ae46-e08499ddf29f" (UID: "44393a30-a46e-4ba0-ae46-e08499ddf29f"). InnerVolumeSpecName "kube-api-access-g65wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.818913 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44393a30-a46e-4ba0-ae46-e08499ddf29f" (UID: "44393a30-a46e-4ba0-ae46-e08499ddf29f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.876409 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g65wt\" (UniqueName: \"kubernetes.io/projected/44393a30-a46e-4ba0-ae46-e08499ddf29f-kube-api-access-g65wt\") on node \"crc\" DevicePath \"\"" Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.876453 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:22:08 crc kubenswrapper[4979]: I1204 12:22:08.876466 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44393a30-a46e-4ba0-ae46-e08499ddf29f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.137868 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rv9jd" event={"ID":"44393a30-a46e-4ba0-ae46-e08499ddf29f","Type":"ContainerDied","Data":"b50698124fc1d696c084baee81f667e4673b94af0d1ba6e6e53973e3ce3382cc"} Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.137896 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rv9jd" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.137916 4979 scope.go:117] "RemoveContainer" containerID="cf225ec7a80620e2281d80aaf696da33977aeab62e585c03c93e163ebfb43f5c" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.138014 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-p4j49" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="registry-server" containerID="cri-o://cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f" gracePeriod=2 Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.159922 4979 scope.go:117] "RemoveContainer" containerID="710062232b1ab1039f70a0ec78f0e357575d0340d09c4f81f43e4705c2122dd1" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.171244 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rv9jd"] Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.178724 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rv9jd"] Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.197682 4979 scope.go:117] "RemoveContainer" containerID="4b89f36ebdee07c018abbf3176ae4243abaf3d483233cb79eb10806dda10404b" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.458249 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.587904 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-utilities\") pod \"618e561d-dcea-4355-931a-c80392f77d66\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.588027 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j52c2\" (UniqueName: \"kubernetes.io/projected/618e561d-dcea-4355-931a-c80392f77d66-kube-api-access-j52c2\") pod \"618e561d-dcea-4355-931a-c80392f77d66\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.588069 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-catalog-content\") pod \"618e561d-dcea-4355-931a-c80392f77d66\" (UID: \"618e561d-dcea-4355-931a-c80392f77d66\") " Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.589612 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-utilities" (OuterVolumeSpecName: "utilities") pod "618e561d-dcea-4355-931a-c80392f77d66" (UID: "618e561d-dcea-4355-931a-c80392f77d66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.593400 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/618e561d-dcea-4355-931a-c80392f77d66-kube-api-access-j52c2" (OuterVolumeSpecName: "kube-api-access-j52c2") pod "618e561d-dcea-4355-931a-c80392f77d66" (UID: "618e561d-dcea-4355-931a-c80392f77d66"). InnerVolumeSpecName "kube-api-access-j52c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.651898 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "618e561d-dcea-4355-931a-c80392f77d66" (UID: "618e561d-dcea-4355-931a-c80392f77d66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.690634 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.690697 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j52c2\" (UniqueName: \"kubernetes.io/projected/618e561d-dcea-4355-931a-c80392f77d66-kube-api-access-j52c2\") on node \"crc\" DevicePath \"\"" Dec 04 12:22:09 crc kubenswrapper[4979]: I1204 12:22:09.690718 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/618e561d-dcea-4355-931a-c80392f77d66-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.150427 4979 generic.go:334] "Generic (PLEG): container finished" podID="618e561d-dcea-4355-931a-c80392f77d66" containerID="cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f" exitCode=0 Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.150509 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p4j49" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.150555 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerDied","Data":"cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f"} Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.150632 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p4j49" event={"ID":"618e561d-dcea-4355-931a-c80392f77d66","Type":"ContainerDied","Data":"ccb7bb3db592371b1f6a2e41b06a2e15471bcda78c1a0ce1d1f5dd92e7720552"} Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.150690 4979 scope.go:117] "RemoveContainer" containerID="cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.191921 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-p4j49"] Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.195094 4979 scope.go:117] "RemoveContainer" containerID="87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.213148 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" path="/var/lib/kubelet/pods/44393a30-a46e-4ba0-ae46-e08499ddf29f/volumes" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.213929 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-p4j49"] Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.228619 4979 scope.go:117] "RemoveContainer" containerID="bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.247221 4979 scope.go:117] "RemoveContainer" containerID="cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f" Dec 04 12:22:10 crc kubenswrapper[4979]: E1204 12:22:10.247753 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f\": container with ID starting with cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f not found: ID does not exist" containerID="cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.247821 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f"} err="failed to get container status \"cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f\": rpc error: code = NotFound desc = could not find container \"cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f\": container with ID starting with cd2e0b83e3ed1722e0fc7f7640b362b8085be7e3641747b9cb3730a28c913e4f not found: ID does not exist" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.247848 4979 scope.go:117] "RemoveContainer" containerID="87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac" Dec 04 12:22:10 crc kubenswrapper[4979]: E1204 12:22:10.248212 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac\": container with ID starting with 87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac not found: ID does not exist" containerID="87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.248233 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac"} err="failed to get container status \"87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac\": rpc error: code = NotFound desc = could not find container \"87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac\": container with ID starting with 87caa51129093eec7e0974732effaadede89c9e4b340cbd1b1551f206193ecac not found: ID does not exist" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.248249 4979 scope.go:117] "RemoveContainer" containerID="bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98" Dec 04 12:22:10 crc kubenswrapper[4979]: E1204 12:22:10.248661 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98\": container with ID starting with bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98 not found: ID does not exist" containerID="bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98" Dec 04 12:22:10 crc kubenswrapper[4979]: I1204 12:22:10.248689 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98"} err="failed to get container status \"bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98\": rpc error: code = NotFound desc = could not find container \"bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98\": container with ID starting with bb23a717380da245384aa518824f2e68205bd0807902fd361e7b1c95393bda98 not found: ID does not exist" Dec 04 12:22:12 crc kubenswrapper[4979]: I1204 12:22:12.214049 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="618e561d-dcea-4355-931a-c80392f77d66" path="/var/lib/kubelet/pods/618e561d-dcea-4355-931a-c80392f77d66/volumes" Dec 04 12:22:19 crc kubenswrapper[4979]: I1204 12:22:19.198604 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:22:19 crc kubenswrapper[4979]: E1204 12:22:19.199371 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:22:32 crc kubenswrapper[4979]: I1204 12:22:32.200089 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:22:32 crc kubenswrapper[4979]: E1204 12:22:32.201448 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:22:43 crc kubenswrapper[4979]: I1204 12:22:43.199424 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:22:43 crc kubenswrapper[4979]: E1204 12:22:43.200457 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:22:57 crc kubenswrapper[4979]: I1204 12:22:57.200148 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:22:57 crc kubenswrapper[4979]: E1204 12:22:57.201105 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:23:12 crc kubenswrapper[4979]: I1204 12:23:12.199338 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:23:12 crc kubenswrapper[4979]: E1204 12:23:12.200026 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:23:24 crc kubenswrapper[4979]: I1204 12:23:24.198550 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:23:24 crc kubenswrapper[4979]: E1204 12:23:24.199207 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:23:39 crc kubenswrapper[4979]: I1204 12:23:39.199168 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:23:39 crc kubenswrapper[4979]: E1204 12:23:39.200125 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:23:53 crc kubenswrapper[4979]: I1204 12:23:53.198703 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:23:53 crc kubenswrapper[4979]: E1204 12:23:53.199516 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:24:06 crc kubenswrapper[4979]: I1204 12:24:06.204031 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:24:06 crc kubenswrapper[4979]: E1204 12:24:06.204858 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:24:21 crc kubenswrapper[4979]: I1204 12:24:21.199041 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:24:21 crc kubenswrapper[4979]: E1204 12:24:21.199724 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:24:34 crc kubenswrapper[4979]: I1204 12:24:34.198904 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:24:34 crc kubenswrapper[4979]: E1204 12:24:34.199774 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:24:48 crc kubenswrapper[4979]: I1204 12:24:48.200087 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:24:48 crc kubenswrapper[4979]: E1204 12:24:48.201088 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:25:01 crc kubenswrapper[4979]: I1204 12:25:01.199116 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:25:01 crc kubenswrapper[4979]: I1204 12:25:01.518191 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"64e38acde448fd486110558dfea8b74626641a7f78e738f916a90f7fd6fadbae"} Dec 04 12:27:28 crc kubenswrapper[4979]: I1204 12:27:28.040730 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:27:28 crc kubenswrapper[4979]: I1204 12:27:28.041377 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:27:58 crc kubenswrapper[4979]: I1204 12:27:58.040767 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:27:58 crc kubenswrapper[4979]: I1204 12:27:58.041216 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.106370 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nrpmn"] Dec 04 12:27:59 crc kubenswrapper[4979]: E1204 12:27:59.107044 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="extract-content" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107060 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="extract-content" Dec 04 12:27:59 crc kubenswrapper[4979]: E1204 12:27:59.107083 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="registry-server" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107093 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="registry-server" Dec 04 12:27:59 crc kubenswrapper[4979]: E1204 12:27:59.107103 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="extract-utilities" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107115 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="extract-utilities" Dec 04 12:27:59 crc kubenswrapper[4979]: E1204 12:27:59.107142 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="extract-content" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107151 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="extract-content" Dec 04 12:27:59 crc kubenswrapper[4979]: E1204 12:27:59.107171 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="extract-utilities" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107180 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="extract-utilities" Dec 04 12:27:59 crc kubenswrapper[4979]: E1204 12:27:59.107197 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="registry-server" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107205 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="registry-server" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107409 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="44393a30-a46e-4ba0-ae46-e08499ddf29f" containerName="registry-server" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.107439 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="618e561d-dcea-4355-931a-c80392f77d66" containerName="registry-server" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.108778 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.124717 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrpmn"] Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.252725 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zph9m\" (UniqueName: \"kubernetes.io/projected/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-kube-api-access-zph9m\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.252792 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-utilities\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.253030 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-catalog-content\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.353964 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zph9m\" (UniqueName: \"kubernetes.io/projected/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-kube-api-access-zph9m\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.354035 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-utilities\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.354101 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-catalog-content\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.354595 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-catalog-content\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.354778 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-utilities\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.379250 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zph9m\" (UniqueName: \"kubernetes.io/projected/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-kube-api-access-zph9m\") pod \"redhat-operators-nrpmn\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.436598 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:27:59 crc kubenswrapper[4979]: I1204 12:27:59.682080 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrpmn"] Dec 04 12:28:00 crc kubenswrapper[4979]: I1204 12:28:00.489047 4979 generic.go:334] "Generic (PLEG): container finished" podID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerID="a954420f4a6f5bd4b2430f0c5cd6abed823bf30723c3c849c38b8f375a07680f" exitCode=0 Dec 04 12:28:00 crc kubenswrapper[4979]: I1204 12:28:00.489158 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerDied","Data":"a954420f4a6f5bd4b2430f0c5cd6abed823bf30723c3c849c38b8f375a07680f"} Dec 04 12:28:00 crc kubenswrapper[4979]: I1204 12:28:00.489387 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerStarted","Data":"9413c47a663836be44a2baedffcc319fed7f1fdb84a09615224a3d1ad098ef1b"} Dec 04 12:28:00 crc kubenswrapper[4979]: I1204 12:28:00.490809 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:28:01 crc kubenswrapper[4979]: I1204 12:28:01.497329 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerStarted","Data":"d771e34d6d74c4e600a62ca4b1e889368efddbfb794c5acef3e983e1ea82b955"} Dec 04 12:28:02 crc kubenswrapper[4979]: I1204 12:28:02.507749 4979 generic.go:334] "Generic (PLEG): container finished" podID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerID="d771e34d6d74c4e600a62ca4b1e889368efddbfb794c5acef3e983e1ea82b955" exitCode=0 Dec 04 12:28:02 crc kubenswrapper[4979]: I1204 12:28:02.507793 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerDied","Data":"d771e34d6d74c4e600a62ca4b1e889368efddbfb794c5acef3e983e1ea82b955"} Dec 04 12:28:04 crc kubenswrapper[4979]: I1204 12:28:04.535952 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerStarted","Data":"2bbb3e5641e7b275b62cc1432e76207bfb90f391703a2db94b3907e276e6ecff"} Dec 04 12:28:04 crc kubenswrapper[4979]: I1204 12:28:04.561267 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nrpmn" podStartSLOduration=2.740039974 podStartE2EDuration="5.561245546s" podCreationTimestamp="2025-12-04 12:27:59 +0000 UTC" firstStartedPulling="2025-12-04 12:28:00.49061207 +0000 UTC m=+2704.764907874" lastFinishedPulling="2025-12-04 12:28:03.311817592 +0000 UTC m=+2707.586113446" observedRunningTime="2025-12-04 12:28:04.558698126 +0000 UTC m=+2708.832993940" watchObservedRunningTime="2025-12-04 12:28:04.561245546 +0000 UTC m=+2708.835541370" Dec 04 12:28:09 crc kubenswrapper[4979]: I1204 12:28:09.436779 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:28:09 crc kubenswrapper[4979]: I1204 12:28:09.437434 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:28:09 crc kubenswrapper[4979]: I1204 12:28:09.525910 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:28:09 crc kubenswrapper[4979]: I1204 12:28:09.629168 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:28:09 crc kubenswrapper[4979]: I1204 12:28:09.768148 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrpmn"] Dec 04 12:28:11 crc kubenswrapper[4979]: I1204 12:28:11.590028 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nrpmn" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="registry-server" containerID="cri-o://2bbb3e5641e7b275b62cc1432e76207bfb90f391703a2db94b3907e276e6ecff" gracePeriod=2 Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.634710 4979 generic.go:334] "Generic (PLEG): container finished" podID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerID="2bbb3e5641e7b275b62cc1432e76207bfb90f391703a2db94b3907e276e6ecff" exitCode=0 Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.634854 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerDied","Data":"2bbb3e5641e7b275b62cc1432e76207bfb90f391703a2db94b3907e276e6ecff"} Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.835391 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.926978 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-catalog-content\") pod \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.927048 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-utilities\") pod \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.927281 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zph9m\" (UniqueName: \"kubernetes.io/projected/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-kube-api-access-zph9m\") pod \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\" (UID: \"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7\") " Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.928729 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-utilities" (OuterVolumeSpecName: "utilities") pod "c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" (UID: "c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:28:15 crc kubenswrapper[4979]: I1204 12:28:15.937044 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-kube-api-access-zph9m" (OuterVolumeSpecName: "kube-api-access-zph9m") pod "c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" (UID: "c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7"). InnerVolumeSpecName "kube-api-access-zph9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.029368 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.029427 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zph9m\" (UniqueName: \"kubernetes.io/projected/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-kube-api-access-zph9m\") on node \"crc\" DevicePath \"\"" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.066879 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" (UID: "c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.130584 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.644154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrpmn" event={"ID":"c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7","Type":"ContainerDied","Data":"9413c47a663836be44a2baedffcc319fed7f1fdb84a09615224a3d1ad098ef1b"} Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.644240 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrpmn" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.645069 4979 scope.go:117] "RemoveContainer" containerID="2bbb3e5641e7b275b62cc1432e76207bfb90f391703a2db94b3907e276e6ecff" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.667750 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrpmn"] Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.675492 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nrpmn"] Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.679022 4979 scope.go:117] "RemoveContainer" containerID="d771e34d6d74c4e600a62ca4b1e889368efddbfb794c5acef3e983e1ea82b955" Dec 04 12:28:16 crc kubenswrapper[4979]: I1204 12:28:16.712492 4979 scope.go:117] "RemoveContainer" containerID="a954420f4a6f5bd4b2430f0c5cd6abed823bf30723c3c849c38b8f375a07680f" Dec 04 12:28:18 crc kubenswrapper[4979]: I1204 12:28:18.213871 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" path="/var/lib/kubelet/pods/c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7/volumes" Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.041702 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.042326 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.042384 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.043080 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"64e38acde448fd486110558dfea8b74626641a7f78e738f916a90f7fd6fadbae"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.043139 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://64e38acde448fd486110558dfea8b74626641a7f78e738f916a90f7fd6fadbae" gracePeriod=600 Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.754639 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="64e38acde448fd486110558dfea8b74626641a7f78e738f916a90f7fd6fadbae" exitCode=0 Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.754761 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"64e38acde448fd486110558dfea8b74626641a7f78e738f916a90f7fd6fadbae"} Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.755057 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d"} Dec 04 12:28:28 crc kubenswrapper[4979]: I1204 12:28:28.755090 4979 scope.go:117] "RemoveContainer" containerID="8d15845b2ac2d9d76b3d05ce0f3d16aa13f42285ccb848179b0918ee091d2f31" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.195944 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6wzvq"] Dec 04 12:28:32 crc kubenswrapper[4979]: E1204 12:28:32.196655 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="extract-content" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.196673 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="extract-content" Dec 04 12:28:32 crc kubenswrapper[4979]: E1204 12:28:32.196699 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="extract-utilities" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.196707 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="extract-utilities" Dec 04 12:28:32 crc kubenswrapper[4979]: E1204 12:28:32.196720 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="registry-server" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.196728 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="registry-server" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.196912 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4cd08b6-5b30-4ba1-bcb4-45bef97f7fa7" containerName="registry-server" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.203679 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.215892 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6wzvq"] Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.276490 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-utilities\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.276907 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-catalog-content\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.276994 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mpnj\" (UniqueName: \"kubernetes.io/projected/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-kube-api-access-7mpnj\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.378187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-catalog-content\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.378271 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mpnj\" (UniqueName: \"kubernetes.io/projected/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-kube-api-access-7mpnj\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.378350 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-utilities\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.378851 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-utilities\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.378849 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-catalog-content\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.398439 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mpnj\" (UniqueName: \"kubernetes.io/projected/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-kube-api-access-7mpnj\") pod \"redhat-marketplace-6wzvq\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.533988 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:32 crc kubenswrapper[4979]: I1204 12:28:32.983698 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6wzvq"] Dec 04 12:28:32 crc kubenswrapper[4979]: W1204 12:28:32.987135 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83daf4df_b406_4926_bbc8_f1fe51c3c6d9.slice/crio-dfc69f3fd3b42f2eec8d5d6616473a5759150ecfd0f466056f2d9e3d449ad68e WatchSource:0}: Error finding container dfc69f3fd3b42f2eec8d5d6616473a5759150ecfd0f466056f2d9e3d449ad68e: Status 404 returned error can't find the container with id dfc69f3fd3b42f2eec8d5d6616473a5759150ecfd0f466056f2d9e3d449ad68e Dec 04 12:28:33 crc kubenswrapper[4979]: I1204 12:28:33.812543 4979 generic.go:334] "Generic (PLEG): container finished" podID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerID="b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2" exitCode=0 Dec 04 12:28:33 crc kubenswrapper[4979]: I1204 12:28:33.812664 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerDied","Data":"b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2"} Dec 04 12:28:33 crc kubenswrapper[4979]: I1204 12:28:33.813011 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerStarted","Data":"dfc69f3fd3b42f2eec8d5d6616473a5759150ecfd0f466056f2d9e3d449ad68e"} Dec 04 12:28:34 crc kubenswrapper[4979]: I1204 12:28:34.820085 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerStarted","Data":"35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288"} Dec 04 12:28:35 crc kubenswrapper[4979]: I1204 12:28:35.835776 4979 generic.go:334] "Generic (PLEG): container finished" podID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerID="35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288" exitCode=0 Dec 04 12:28:35 crc kubenswrapper[4979]: I1204 12:28:35.835969 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerDied","Data":"35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288"} Dec 04 12:28:36 crc kubenswrapper[4979]: I1204 12:28:36.851691 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerStarted","Data":"c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3"} Dec 04 12:28:36 crc kubenswrapper[4979]: I1204 12:28:36.880341 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6wzvq" podStartSLOduration=2.416845291 podStartE2EDuration="4.880315366s" podCreationTimestamp="2025-12-04 12:28:32 +0000 UTC" firstStartedPulling="2025-12-04 12:28:33.81574722 +0000 UTC m=+2738.090043034" lastFinishedPulling="2025-12-04 12:28:36.279217305 +0000 UTC m=+2740.553513109" observedRunningTime="2025-12-04 12:28:36.87240717 +0000 UTC m=+2741.146702974" watchObservedRunningTime="2025-12-04 12:28:36.880315366 +0000 UTC m=+2741.154611170" Dec 04 12:28:42 crc kubenswrapper[4979]: I1204 12:28:42.534872 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:42 crc kubenswrapper[4979]: I1204 12:28:42.535272 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:42 crc kubenswrapper[4979]: I1204 12:28:42.611941 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:43 crc kubenswrapper[4979]: I1204 12:28:43.478936 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:43 crc kubenswrapper[4979]: I1204 12:28:43.535146 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6wzvq"] Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.438695 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6wzvq" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="registry-server" containerID="cri-o://c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3" gracePeriod=2 Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.839811 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.963505 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-catalog-content\") pod \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.963614 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mpnj\" (UniqueName: \"kubernetes.io/projected/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-kube-api-access-7mpnj\") pod \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.963716 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-utilities\") pod \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\" (UID: \"83daf4df-b406-4926-bbc8-f1fe51c3c6d9\") " Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.965272 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-utilities" (OuterVolumeSpecName: "utilities") pod "83daf4df-b406-4926-bbc8-f1fe51c3c6d9" (UID: "83daf4df-b406-4926-bbc8-f1fe51c3c6d9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.969029 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-kube-api-access-7mpnj" (OuterVolumeSpecName: "kube-api-access-7mpnj") pod "83daf4df-b406-4926-bbc8-f1fe51c3c6d9" (UID: "83daf4df-b406-4926-bbc8-f1fe51c3c6d9"). InnerVolumeSpecName "kube-api-access-7mpnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:28:45 crc kubenswrapper[4979]: I1204 12:28:45.998431 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83daf4df-b406-4926-bbc8-f1fe51c3c6d9" (UID: "83daf4df-b406-4926-bbc8-f1fe51c3c6d9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.065839 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.065876 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mpnj\" (UniqueName: \"kubernetes.io/projected/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-kube-api-access-7mpnj\") on node \"crc\" DevicePath \"\"" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.065888 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83daf4df-b406-4926-bbc8-f1fe51c3c6d9-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.451000 4979 generic.go:334] "Generic (PLEG): container finished" podID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerID="c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3" exitCode=0 Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.451134 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6wzvq" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.451166 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerDied","Data":"c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3"} Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.451350 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6wzvq" event={"ID":"83daf4df-b406-4926-bbc8-f1fe51c3c6d9","Type":"ContainerDied","Data":"dfc69f3fd3b42f2eec8d5d6616473a5759150ecfd0f466056f2d9e3d449ad68e"} Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.451422 4979 scope.go:117] "RemoveContainer" containerID="c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.478974 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6wzvq"] Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.479187 4979 scope.go:117] "RemoveContainer" containerID="35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.484549 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6wzvq"] Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.507915 4979 scope.go:117] "RemoveContainer" containerID="b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.533630 4979 scope.go:117] "RemoveContainer" containerID="c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3" Dec 04 12:28:46 crc kubenswrapper[4979]: E1204 12:28:46.534249 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3\": container with ID starting with c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3 not found: ID does not exist" containerID="c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.534288 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3"} err="failed to get container status \"c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3\": rpc error: code = NotFound desc = could not find container \"c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3\": container with ID starting with c258af17d6316b0f25b196f2f9942b69a02b1cc7082fa0a0d9926a7af1a137b3 not found: ID does not exist" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.534330 4979 scope.go:117] "RemoveContainer" containerID="35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288" Dec 04 12:28:46 crc kubenswrapper[4979]: E1204 12:28:46.534660 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288\": container with ID starting with 35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288 not found: ID does not exist" containerID="35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.534707 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288"} err="failed to get container status \"35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288\": rpc error: code = NotFound desc = could not find container \"35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288\": container with ID starting with 35d78d9e150235792128e1673d068b986578eeb0b645ec9baa3420af568bc288 not found: ID does not exist" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.534762 4979 scope.go:117] "RemoveContainer" containerID="b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2" Dec 04 12:28:46 crc kubenswrapper[4979]: E1204 12:28:46.535529 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2\": container with ID starting with b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2 not found: ID does not exist" containerID="b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2" Dec 04 12:28:46 crc kubenswrapper[4979]: I1204 12:28:46.535557 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2"} err="failed to get container status \"b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2\": rpc error: code = NotFound desc = could not find container \"b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2\": container with ID starting with b9db637a4033cd272291d95b787bada83856471f2b048a406294d9d2e4d313a2 not found: ID does not exist" Dec 04 12:28:48 crc kubenswrapper[4979]: I1204 12:28:48.208849 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" path="/var/lib/kubelet/pods/83daf4df-b406-4926-bbc8-f1fe51c3c6d9/volumes" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.157875 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg"] Dec 04 12:30:00 crc kubenswrapper[4979]: E1204 12:30:00.159209 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="extract-utilities" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.159245 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="extract-utilities" Dec 04 12:30:00 crc kubenswrapper[4979]: E1204 12:30:00.159280 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="extract-content" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.159356 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="extract-content" Dec 04 12:30:00 crc kubenswrapper[4979]: E1204 12:30:00.159399 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="registry-server" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.159413 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="registry-server" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.159745 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="83daf4df-b406-4926-bbc8-f1fe51c3c6d9" containerName="registry-server" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.160699 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.163165 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.164984 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.178991 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg"] Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.299801 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a6d345d-320a-4625-88dc-17507072c856-secret-volume\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.300647 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp8qg\" (UniqueName: \"kubernetes.io/projected/4a6d345d-320a-4625-88dc-17507072c856-kube-api-access-tp8qg\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.300733 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a6d345d-320a-4625-88dc-17507072c856-config-volume\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.402581 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp8qg\" (UniqueName: \"kubernetes.io/projected/4a6d345d-320a-4625-88dc-17507072c856-kube-api-access-tp8qg\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.402650 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a6d345d-320a-4625-88dc-17507072c856-config-volume\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.402729 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a6d345d-320a-4625-88dc-17507072c856-secret-volume\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.403549 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a6d345d-320a-4625-88dc-17507072c856-config-volume\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.408798 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a6d345d-320a-4625-88dc-17507072c856-secret-volume\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.418189 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp8qg\" (UniqueName: \"kubernetes.io/projected/4a6d345d-320a-4625-88dc-17507072c856-kube-api-access-tp8qg\") pod \"collect-profiles-29414190-mqhlg\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.481778 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:00 crc kubenswrapper[4979]: I1204 12:30:00.887616 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg"] Dec 04 12:30:01 crc kubenswrapper[4979]: I1204 12:30:01.064095 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" event={"ID":"4a6d345d-320a-4625-88dc-17507072c856","Type":"ContainerStarted","Data":"d898af5d009a09db135552403dd9b833252568ae039d3a599335f6d45551766d"} Dec 04 12:30:02 crc kubenswrapper[4979]: I1204 12:30:02.074482 4979 generic.go:334] "Generic (PLEG): container finished" podID="4a6d345d-320a-4625-88dc-17507072c856" containerID="65274224dc2010c1de99ebd5b6d6eb2b84e48e3829a3444d156c4d22b5397cca" exitCode=0 Dec 04 12:30:02 crc kubenswrapper[4979]: I1204 12:30:02.074581 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" event={"ID":"4a6d345d-320a-4625-88dc-17507072c856","Type":"ContainerDied","Data":"65274224dc2010c1de99ebd5b6d6eb2b84e48e3829a3444d156c4d22b5397cca"} Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.366257 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.457390 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a6d345d-320a-4625-88dc-17507072c856-config-volume\") pod \"4a6d345d-320a-4625-88dc-17507072c856\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.457483 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp8qg\" (UniqueName: \"kubernetes.io/projected/4a6d345d-320a-4625-88dc-17507072c856-kube-api-access-tp8qg\") pod \"4a6d345d-320a-4625-88dc-17507072c856\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.457586 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a6d345d-320a-4625-88dc-17507072c856-secret-volume\") pod \"4a6d345d-320a-4625-88dc-17507072c856\" (UID: \"4a6d345d-320a-4625-88dc-17507072c856\") " Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.458888 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a6d345d-320a-4625-88dc-17507072c856-config-volume" (OuterVolumeSpecName: "config-volume") pod "4a6d345d-320a-4625-88dc-17507072c856" (UID: "4a6d345d-320a-4625-88dc-17507072c856"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.461938 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a6d345d-320a-4625-88dc-17507072c856-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4a6d345d-320a-4625-88dc-17507072c856" (UID: "4a6d345d-320a-4625-88dc-17507072c856"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.463405 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a6d345d-320a-4625-88dc-17507072c856-kube-api-access-tp8qg" (OuterVolumeSpecName: "kube-api-access-tp8qg") pod "4a6d345d-320a-4625-88dc-17507072c856" (UID: "4a6d345d-320a-4625-88dc-17507072c856"). InnerVolumeSpecName "kube-api-access-tp8qg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.558961 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4a6d345d-320a-4625-88dc-17507072c856-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.559006 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp8qg\" (UniqueName: \"kubernetes.io/projected/4a6d345d-320a-4625-88dc-17507072c856-kube-api-access-tp8qg\") on node \"crc\" DevicePath \"\"" Dec 04 12:30:03 crc kubenswrapper[4979]: I1204 12:30:03.559023 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4a6d345d-320a-4625-88dc-17507072c856-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:30:04 crc kubenswrapper[4979]: I1204 12:30:04.092685 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" event={"ID":"4a6d345d-320a-4625-88dc-17507072c856","Type":"ContainerDied","Data":"d898af5d009a09db135552403dd9b833252568ae039d3a599335f6d45551766d"} Dec 04 12:30:04 crc kubenswrapper[4979]: I1204 12:30:04.092727 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d898af5d009a09db135552403dd9b833252568ae039d3a599335f6d45551766d" Dec 04 12:30:04 crc kubenswrapper[4979]: I1204 12:30:04.092803 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg" Dec 04 12:30:04 crc kubenswrapper[4979]: I1204 12:30:04.442315 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc"] Dec 04 12:30:04 crc kubenswrapper[4979]: I1204 12:30:04.447754 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414145-4glpc"] Dec 04 12:30:06 crc kubenswrapper[4979]: I1204 12:30:06.210440 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5056ac5a-1bc7-4298-9417-995c16c28ec1" path="/var/lib/kubelet/pods/5056ac5a-1bc7-4298-9417-995c16c28ec1/volumes" Dec 04 12:30:28 crc kubenswrapper[4979]: I1204 12:30:28.040963 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:30:28 crc kubenswrapper[4979]: I1204 12:30:28.041726 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:30:58 crc kubenswrapper[4979]: I1204 12:30:58.041550 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:30:58 crc kubenswrapper[4979]: I1204 12:30:58.042077 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:31:05 crc kubenswrapper[4979]: I1204 12:31:05.029990 4979 scope.go:117] "RemoveContainer" containerID="8b201fdda6f56d0c79110919245a41b83c144ddc3c0dab7efb19a389dc3e71d6" Dec 04 12:31:28 crc kubenswrapper[4979]: I1204 12:31:28.041168 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:31:28 crc kubenswrapper[4979]: I1204 12:31:28.041911 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:31:28 crc kubenswrapper[4979]: I1204 12:31:28.041978 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:31:28 crc kubenswrapper[4979]: I1204 12:31:28.042829 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:31:28 crc kubenswrapper[4979]: I1204 12:31:28.042912 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" gracePeriod=600 Dec 04 12:31:29 crc kubenswrapper[4979]: I1204 12:31:29.780343 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" exitCode=0 Dec 04 12:31:29 crc kubenswrapper[4979]: I1204 12:31:29.780388 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d"} Dec 04 12:31:29 crc kubenswrapper[4979]: I1204 12:31:29.780790 4979 scope.go:117] "RemoveContainer" containerID="64e38acde448fd486110558dfea8b74626641a7f78e738f916a90f7fd6fadbae" Dec 04 12:31:31 crc kubenswrapper[4979]: E1204 12:31:31.172146 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:31:31 crc kubenswrapper[4979]: I1204 12:31:31.802027 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:31:31 crc kubenswrapper[4979]: E1204 12:31:31.802322 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:31:47 crc kubenswrapper[4979]: I1204 12:31:47.198643 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:31:47 crc kubenswrapper[4979]: E1204 12:31:47.199523 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:32:00 crc kubenswrapper[4979]: I1204 12:32:00.199434 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:32:00 crc kubenswrapper[4979]: E1204 12:32:00.200005 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:32:13 crc kubenswrapper[4979]: I1204 12:32:13.199401 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:32:13 crc kubenswrapper[4979]: E1204 12:32:13.199961 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:32:24 crc kubenswrapper[4979]: I1204 12:32:24.199945 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:32:24 crc kubenswrapper[4979]: E1204 12:32:24.200916 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:32:39 crc kubenswrapper[4979]: I1204 12:32:39.199684 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:32:39 crc kubenswrapper[4979]: E1204 12:32:39.200563 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:32:51 crc kubenswrapper[4979]: I1204 12:32:51.199295 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:32:51 crc kubenswrapper[4979]: E1204 12:32:51.200516 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:33:04 crc kubenswrapper[4979]: I1204 12:33:04.198856 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:33:04 crc kubenswrapper[4979]: E1204 12:33:04.199683 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:33:15 crc kubenswrapper[4979]: I1204 12:33:15.199394 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:33:15 crc kubenswrapper[4979]: E1204 12:33:15.200637 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.055546 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-plh99"] Dec 04 12:33:16 crc kubenswrapper[4979]: E1204 12:33:16.056258 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a6d345d-320a-4625-88dc-17507072c856" containerName="collect-profiles" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.056273 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a6d345d-320a-4625-88dc-17507072c856" containerName="collect-profiles" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.056451 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a6d345d-320a-4625-88dc-17507072c856" containerName="collect-profiles" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.057580 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.078049 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plh99"] Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.255690 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-catalog-content\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.255766 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdlkn\" (UniqueName: \"kubernetes.io/projected/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-kube-api-access-zdlkn\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.255910 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-utilities\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.356943 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-catalog-content\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.357026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdlkn\" (UniqueName: \"kubernetes.io/projected/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-kube-api-access-zdlkn\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.357072 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-utilities\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.357679 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-catalog-content\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.357703 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-utilities\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.380224 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdlkn\" (UniqueName: \"kubernetes.io/projected/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-kube-api-access-zdlkn\") pod \"certified-operators-plh99\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:16 crc kubenswrapper[4979]: I1204 12:33:16.675404 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:17 crc kubenswrapper[4979]: I1204 12:33:17.065443 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-plh99"] Dec 04 12:33:17 crc kubenswrapper[4979]: I1204 12:33:17.603744 4979 generic.go:334] "Generic (PLEG): container finished" podID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerID="c911948d6bc3d3ad6a72141b2006cfaecad1a02d7b47d9298d2266e48652980b" exitCode=0 Dec 04 12:33:17 crc kubenswrapper[4979]: I1204 12:33:17.603851 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerDied","Data":"c911948d6bc3d3ad6a72141b2006cfaecad1a02d7b47d9298d2266e48652980b"} Dec 04 12:33:17 crc kubenswrapper[4979]: I1204 12:33:17.604078 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerStarted","Data":"89f9c4e921eccda77ee9f1acdbad7e26fd670f79ce941446ec20b2fbc8bea984"} Dec 04 12:33:17 crc kubenswrapper[4979]: I1204 12:33:17.608171 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:33:18 crc kubenswrapper[4979]: I1204 12:33:18.613360 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerStarted","Data":"ac4c53027b11381ee734b9b26d0f850d394bc9e89c49b24e1da893cbd642a6a3"} Dec 04 12:33:19 crc kubenswrapper[4979]: I1204 12:33:19.621340 4979 generic.go:334] "Generic (PLEG): container finished" podID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerID="ac4c53027b11381ee734b9b26d0f850d394bc9e89c49b24e1da893cbd642a6a3" exitCode=0 Dec 04 12:33:19 crc kubenswrapper[4979]: I1204 12:33:19.621440 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerDied","Data":"ac4c53027b11381ee734b9b26d0f850d394bc9e89c49b24e1da893cbd642a6a3"} Dec 04 12:33:20 crc kubenswrapper[4979]: I1204 12:33:20.633268 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerStarted","Data":"50b1b6ff31c20f6d68753594ab42da3d1a9d972e58c199399b9f53fa6c56fcf9"} Dec 04 12:33:20 crc kubenswrapper[4979]: I1204 12:33:20.656164 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-plh99" podStartSLOduration=1.97534308 podStartE2EDuration="4.656140396s" podCreationTimestamp="2025-12-04 12:33:16 +0000 UTC" firstStartedPulling="2025-12-04 12:33:17.607821807 +0000 UTC m=+3021.882117621" lastFinishedPulling="2025-12-04 12:33:20.288619133 +0000 UTC m=+3024.562914937" observedRunningTime="2025-12-04 12:33:20.651125479 +0000 UTC m=+3024.925421333" watchObservedRunningTime="2025-12-04 12:33:20.656140396 +0000 UTC m=+3024.930436200" Dec 04 12:33:26 crc kubenswrapper[4979]: I1204 12:33:26.676180 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:26 crc kubenswrapper[4979]: I1204 12:33:26.676854 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:26 crc kubenswrapper[4979]: I1204 12:33:26.716392 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:27 crc kubenswrapper[4979]: I1204 12:33:27.723942 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:27 crc kubenswrapper[4979]: I1204 12:33:27.778359 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-plh99"] Dec 04 12:33:29 crc kubenswrapper[4979]: I1204 12:33:29.199329 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:33:29 crc kubenswrapper[4979]: E1204 12:33:29.199555 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:33:29 crc kubenswrapper[4979]: I1204 12:33:29.699148 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-plh99" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="registry-server" containerID="cri-o://50b1b6ff31c20f6d68753594ab42da3d1a9d972e58c199399b9f53fa6c56fcf9" gracePeriod=2 Dec 04 12:33:30 crc kubenswrapper[4979]: I1204 12:33:30.711727 4979 generic.go:334] "Generic (PLEG): container finished" podID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerID="50b1b6ff31c20f6d68753594ab42da3d1a9d972e58c199399b9f53fa6c56fcf9" exitCode=0 Dec 04 12:33:30 crc kubenswrapper[4979]: I1204 12:33:30.711793 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerDied","Data":"50b1b6ff31c20f6d68753594ab42da3d1a9d972e58c199399b9f53fa6c56fcf9"} Dec 04 12:33:30 crc kubenswrapper[4979]: E1204 12:33:30.716093 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb84e7dcc_3bde_4e0c_a612_0ef7fe930956.slice/crio-conmon-50b1b6ff31c20f6d68753594ab42da3d1a9d972e58c199399b9f53fa6c56fcf9.scope\": RecentStats: unable to find data in memory cache]" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.208050 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.361054 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-utilities\") pod \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.361143 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-catalog-content\") pod \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.361375 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdlkn\" (UniqueName: \"kubernetes.io/projected/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-kube-api-access-zdlkn\") pod \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\" (UID: \"b84e7dcc-3bde-4e0c-a612-0ef7fe930956\") " Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.362198 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-utilities" (OuterVolumeSpecName: "utilities") pod "b84e7dcc-3bde-4e0c-a612-0ef7fe930956" (UID: "b84e7dcc-3bde-4e0c-a612-0ef7fe930956"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.366647 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-kube-api-access-zdlkn" (OuterVolumeSpecName: "kube-api-access-zdlkn") pod "b84e7dcc-3bde-4e0c-a612-0ef7fe930956" (UID: "b84e7dcc-3bde-4e0c-a612-0ef7fe930956"). InnerVolumeSpecName "kube-api-access-zdlkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.409232 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b84e7dcc-3bde-4e0c-a612-0ef7fe930956" (UID: "b84e7dcc-3bde-4e0c-a612-0ef7fe930956"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.462507 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdlkn\" (UniqueName: \"kubernetes.io/projected/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-kube-api-access-zdlkn\") on node \"crc\" DevicePath \"\"" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.462547 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.462563 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b84e7dcc-3bde-4e0c-a612-0ef7fe930956-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.722346 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-plh99" event={"ID":"b84e7dcc-3bde-4e0c-a612-0ef7fe930956","Type":"ContainerDied","Data":"89f9c4e921eccda77ee9f1acdbad7e26fd670f79ce941446ec20b2fbc8bea984"} Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.722408 4979 scope.go:117] "RemoveContainer" containerID="50b1b6ff31c20f6d68753594ab42da3d1a9d972e58c199399b9f53fa6c56fcf9" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.722410 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-plh99" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.742665 4979 scope.go:117] "RemoveContainer" containerID="ac4c53027b11381ee734b9b26d0f850d394bc9e89c49b24e1da893cbd642a6a3" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.766281 4979 scope.go:117] "RemoveContainer" containerID="c911948d6bc3d3ad6a72141b2006cfaecad1a02d7b47d9298d2266e48652980b" Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.768045 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-plh99"] Dec 04 12:33:31 crc kubenswrapper[4979]: I1204 12:33:31.774177 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-plh99"] Dec 04 12:33:32 crc kubenswrapper[4979]: I1204 12:33:32.214720 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" path="/var/lib/kubelet/pods/b84e7dcc-3bde-4e0c-a612-0ef7fe930956/volumes" Dec 04 12:33:43 crc kubenswrapper[4979]: I1204 12:33:43.199288 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:33:43 crc kubenswrapper[4979]: E1204 12:33:43.200180 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:33:54 crc kubenswrapper[4979]: I1204 12:33:54.199676 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:33:54 crc kubenswrapper[4979]: E1204 12:33:54.200773 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:34:07 crc kubenswrapper[4979]: I1204 12:34:07.198956 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:34:07 crc kubenswrapper[4979]: E1204 12:34:07.199675 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:34:21 crc kubenswrapper[4979]: I1204 12:34:21.200638 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:34:21 crc kubenswrapper[4979]: E1204 12:34:21.202173 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:34:35 crc kubenswrapper[4979]: I1204 12:34:35.199613 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:34:35 crc kubenswrapper[4979]: E1204 12:34:35.200661 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:34:48 crc kubenswrapper[4979]: I1204 12:34:48.199928 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:34:48 crc kubenswrapper[4979]: E1204 12:34:48.202518 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:35:03 crc kubenswrapper[4979]: I1204 12:35:03.199690 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:35:03 crc kubenswrapper[4979]: E1204 12:35:03.200626 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:35:14 crc kubenswrapper[4979]: I1204 12:35:14.200169 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:35:14 crc kubenswrapper[4979]: E1204 12:35:14.201092 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:35:25 crc kubenswrapper[4979]: I1204 12:35:25.199410 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:35:25 crc kubenswrapper[4979]: E1204 12:35:25.200063 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:35:40 crc kubenswrapper[4979]: I1204 12:35:40.199727 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:35:40 crc kubenswrapper[4979]: E1204 12:35:40.201054 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:35:54 crc kubenswrapper[4979]: I1204 12:35:54.199014 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:35:54 crc kubenswrapper[4979]: E1204 12:35:54.199871 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:36:05 crc kubenswrapper[4979]: I1204 12:36:05.199502 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:36:05 crc kubenswrapper[4979]: E1204 12:36:05.200439 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:36:16 crc kubenswrapper[4979]: I1204 12:36:16.207626 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:36:16 crc kubenswrapper[4979]: E1204 12:36:16.208539 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:36:28 crc kubenswrapper[4979]: I1204 12:36:28.199380 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:36:28 crc kubenswrapper[4979]: E1204 12:36:28.200337 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:36:40 crc kubenswrapper[4979]: I1204 12:36:40.199116 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:36:41 crc kubenswrapper[4979]: I1204 12:36:41.173616 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f"} Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.590174 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4xlng"] Dec 04 12:37:23 crc kubenswrapper[4979]: E1204 12:37:23.591066 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="extract-content" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.591083 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="extract-content" Dec 04 12:37:23 crc kubenswrapper[4979]: E1204 12:37:23.591101 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="registry-server" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.591108 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="registry-server" Dec 04 12:37:23 crc kubenswrapper[4979]: E1204 12:37:23.591117 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="extract-utilities" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.591125 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="extract-utilities" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.591328 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b84e7dcc-3bde-4e0c-a612-0ef7fe930956" containerName="registry-server" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.592566 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.606395 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4xlng"] Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.738266 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75tnp\" (UniqueName: \"kubernetes.io/projected/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-kube-api-access-75tnp\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.738335 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-utilities\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.738375 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-catalog-content\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.840190 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75tnp\" (UniqueName: \"kubernetes.io/projected/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-kube-api-access-75tnp\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.840258 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-utilities\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.840311 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-catalog-content\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.840907 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-utilities\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.840918 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-catalog-content\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.861993 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75tnp\" (UniqueName: \"kubernetes.io/projected/5cc37280-3e5a-41d6-ae14-0bd9fc40ae42-kube-api-access-75tnp\") pod \"community-operators-4xlng\" (UID: \"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42\") " pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:23 crc kubenswrapper[4979]: I1204 12:37:23.918236 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:24 crc kubenswrapper[4979]: I1204 12:37:24.352482 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4xlng"] Dec 04 12:37:24 crc kubenswrapper[4979]: I1204 12:37:24.518779 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xlng" event={"ID":"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42","Type":"ContainerStarted","Data":"40f3a00246021cfac759218ce4568c721aeeb6e41cb05985ef7d012aa7ac62d7"} Dec 04 12:37:25 crc kubenswrapper[4979]: I1204 12:37:25.528141 4979 generic.go:334] "Generic (PLEG): container finished" podID="5cc37280-3e5a-41d6-ae14-0bd9fc40ae42" containerID="ccce3db03789b35aab239596663caae44b6cf7379ada6e2cabce5bc4707d4866" exitCode=0 Dec 04 12:37:25 crc kubenswrapper[4979]: I1204 12:37:25.528203 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xlng" event={"ID":"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42","Type":"ContainerDied","Data":"ccce3db03789b35aab239596663caae44b6cf7379ada6e2cabce5bc4707d4866"} Dec 04 12:37:29 crc kubenswrapper[4979]: I1204 12:37:29.562375 4979 generic.go:334] "Generic (PLEG): container finished" podID="5cc37280-3e5a-41d6-ae14-0bd9fc40ae42" containerID="9db086c54db73d06e98614209dbb34ac52af39bdb20171a781f7028fedb728a5" exitCode=0 Dec 04 12:37:29 crc kubenswrapper[4979]: I1204 12:37:29.562607 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xlng" event={"ID":"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42","Type":"ContainerDied","Data":"9db086c54db73d06e98614209dbb34ac52af39bdb20171a781f7028fedb728a5"} Dec 04 12:37:30 crc kubenswrapper[4979]: I1204 12:37:30.572515 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4xlng" event={"ID":"5cc37280-3e5a-41d6-ae14-0bd9fc40ae42","Type":"ContainerStarted","Data":"8d2b50d8256583a28e06340ad999b28ea239697fd274abfa6b271286a6415d24"} Dec 04 12:37:30 crc kubenswrapper[4979]: I1204 12:37:30.590683 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4xlng" podStartSLOduration=3.034448444 podStartE2EDuration="7.59066633s" podCreationTimestamp="2025-12-04 12:37:23 +0000 UTC" firstStartedPulling="2025-12-04 12:37:25.530423761 +0000 UTC m=+3269.804719605" lastFinishedPulling="2025-12-04 12:37:30.086641687 +0000 UTC m=+3274.360937491" observedRunningTime="2025-12-04 12:37:30.586367694 +0000 UTC m=+3274.860663508" watchObservedRunningTime="2025-12-04 12:37:30.59066633 +0000 UTC m=+3274.864962134" Dec 04 12:37:33 crc kubenswrapper[4979]: I1204 12:37:33.919496 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:33 crc kubenswrapper[4979]: I1204 12:37:33.920424 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:33 crc kubenswrapper[4979]: I1204 12:37:33.966024 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:43 crc kubenswrapper[4979]: I1204 12:37:43.972414 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4xlng" Dec 04 12:37:44 crc kubenswrapper[4979]: I1204 12:37:44.041632 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4xlng"] Dec 04 12:37:44 crc kubenswrapper[4979]: I1204 12:37:44.080154 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7n8l"] Dec 04 12:37:44 crc kubenswrapper[4979]: I1204 12:37:44.080434 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d7n8l" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="registry-server" containerID="cri-o://22936e3142b9b6f3558f62c499f6164b3c0a8e615ffe30383fc5ef0368e42094" gracePeriod=2 Dec 04 12:37:44 crc kubenswrapper[4979]: I1204 12:37:44.670465 4979 generic.go:334] "Generic (PLEG): container finished" podID="28921e2c-1159-4b88-bfbf-0214efef7819" containerID="22936e3142b9b6f3558f62c499f6164b3c0a8e615ffe30383fc5ef0368e42094" exitCode=0 Dec 04 12:37:44 crc kubenswrapper[4979]: I1204 12:37:44.670682 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerDied","Data":"22936e3142b9b6f3558f62c499f6164b3c0a8e615ffe30383fc5ef0368e42094"} Dec 04 12:37:44 crc kubenswrapper[4979]: I1204 12:37:44.982259 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.046378 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-catalog-content\") pod \"28921e2c-1159-4b88-bfbf-0214efef7819\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.046477 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ffgl\" (UniqueName: \"kubernetes.io/projected/28921e2c-1159-4b88-bfbf-0214efef7819-kube-api-access-6ffgl\") pod \"28921e2c-1159-4b88-bfbf-0214efef7819\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.046522 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-utilities\") pod \"28921e2c-1159-4b88-bfbf-0214efef7819\" (UID: \"28921e2c-1159-4b88-bfbf-0214efef7819\") " Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.047342 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-utilities" (OuterVolumeSpecName: "utilities") pod "28921e2c-1159-4b88-bfbf-0214efef7819" (UID: "28921e2c-1159-4b88-bfbf-0214efef7819"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.052964 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28921e2c-1159-4b88-bfbf-0214efef7819-kube-api-access-6ffgl" (OuterVolumeSpecName: "kube-api-access-6ffgl") pod "28921e2c-1159-4b88-bfbf-0214efef7819" (UID: "28921e2c-1159-4b88-bfbf-0214efef7819"). InnerVolumeSpecName "kube-api-access-6ffgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.102729 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28921e2c-1159-4b88-bfbf-0214efef7819" (UID: "28921e2c-1159-4b88-bfbf-0214efef7819"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.148112 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.148485 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ffgl\" (UniqueName: \"kubernetes.io/projected/28921e2c-1159-4b88-bfbf-0214efef7819-kube-api-access-6ffgl\") on node \"crc\" DevicePath \"\"" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.148634 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28921e2c-1159-4b88-bfbf-0214efef7819-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.684762 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d7n8l" event={"ID":"28921e2c-1159-4b88-bfbf-0214efef7819","Type":"ContainerDied","Data":"91329605ed1b4c95e32531b6ce834f968a09ac22c4a116797c82d4f8ee89e679"} Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.685125 4979 scope.go:117] "RemoveContainer" containerID="22936e3142b9b6f3558f62c499f6164b3c0a8e615ffe30383fc5ef0368e42094" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.684901 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d7n8l" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.710195 4979 scope.go:117] "RemoveContainer" containerID="b67d4e615023233376bf96346977e618cfd637c0cf373f966265f9cd69781dfc" Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.729736 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d7n8l"] Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.740650 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d7n8l"] Dec 04 12:37:45 crc kubenswrapper[4979]: I1204 12:37:45.746083 4979 scope.go:117] "RemoveContainer" containerID="8c155757f1feb7bdbb352e5ee02882f2c51bcc8e47d840ae3653adaaa38a5108" Dec 04 12:37:46 crc kubenswrapper[4979]: I1204 12:37:46.212942 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" path="/var/lib/kubelet/pods/28921e2c-1159-4b88-bfbf-0214efef7819/volumes" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.031002 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-q4dl2"] Dec 04 12:38:25 crc kubenswrapper[4979]: E1204 12:38:25.031907 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="extract-content" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.031922 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="extract-content" Dec 04 12:38:25 crc kubenswrapper[4979]: E1204 12:38:25.031948 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="extract-utilities" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.031956 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="extract-utilities" Dec 04 12:38:25 crc kubenswrapper[4979]: E1204 12:38:25.031973 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="registry-server" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.031983 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="registry-server" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.032170 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="28921e2c-1159-4b88-bfbf-0214efef7819" containerName="registry-server" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.033343 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.048548 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q4dl2"] Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.132869 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g49ht\" (UniqueName: \"kubernetes.io/projected/ce253cac-4481-43be-8091-d15072eaf098-kube-api-access-g49ht\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.133166 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-utilities\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.133312 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-catalog-content\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.234221 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-catalog-content\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.234667 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g49ht\" (UniqueName: \"kubernetes.io/projected/ce253cac-4481-43be-8091-d15072eaf098-kube-api-access-g49ht\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.234779 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-utilities\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.235781 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-catalog-content\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.236103 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-utilities\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.255088 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g49ht\" (UniqueName: \"kubernetes.io/projected/ce253cac-4481-43be-8091-d15072eaf098-kube-api-access-g49ht\") pod \"redhat-operators-q4dl2\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.352166 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:25 crc kubenswrapper[4979]: I1204 12:38:25.774829 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-q4dl2"] Dec 04 12:38:26 crc kubenswrapper[4979]: I1204 12:38:26.000123 4979 generic.go:334] "Generic (PLEG): container finished" podID="ce253cac-4481-43be-8091-d15072eaf098" containerID="ead85f559c6aaed4face0b3fed52dd5281f2c6cf1b7dccafdecaf7ea2885305b" exitCode=0 Dec 04 12:38:26 crc kubenswrapper[4979]: I1204 12:38:26.000171 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerDied","Data":"ead85f559c6aaed4face0b3fed52dd5281f2c6cf1b7dccafdecaf7ea2885305b"} Dec 04 12:38:26 crc kubenswrapper[4979]: I1204 12:38:26.000200 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerStarted","Data":"fef2668763dbe1d5edf1c06eb4a84689aa1db15e8c96bba50707ba3134f468e4"} Dec 04 12:38:26 crc kubenswrapper[4979]: I1204 12:38:26.001800 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:38:27 crc kubenswrapper[4979]: I1204 12:38:27.009362 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerStarted","Data":"136c417568ee24e7a878196a8ebcf854e488f0239d74527dfa345a11bcddb3a8"} Dec 04 12:38:28 crc kubenswrapper[4979]: I1204 12:38:28.023488 4979 generic.go:334] "Generic (PLEG): container finished" podID="ce253cac-4481-43be-8091-d15072eaf098" containerID="136c417568ee24e7a878196a8ebcf854e488f0239d74527dfa345a11bcddb3a8" exitCode=0 Dec 04 12:38:28 crc kubenswrapper[4979]: I1204 12:38:28.023563 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerDied","Data":"136c417568ee24e7a878196a8ebcf854e488f0239d74527dfa345a11bcddb3a8"} Dec 04 12:38:29 crc kubenswrapper[4979]: I1204 12:38:29.033357 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerStarted","Data":"5eb32f63726425d8fca1025ea619fc8205e9f9e637480bf5a10a0b9300c4b28b"} Dec 04 12:38:29 crc kubenswrapper[4979]: I1204 12:38:29.057505 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-q4dl2" podStartSLOduration=1.65477596 podStartE2EDuration="4.057487426s" podCreationTimestamp="2025-12-04 12:38:25 +0000 UTC" firstStartedPulling="2025-12-04 12:38:26.001502719 +0000 UTC m=+3330.275798523" lastFinishedPulling="2025-12-04 12:38:28.404214185 +0000 UTC m=+3332.678509989" observedRunningTime="2025-12-04 12:38:29.052270394 +0000 UTC m=+3333.326566218" watchObservedRunningTime="2025-12-04 12:38:29.057487426 +0000 UTC m=+3333.331783230" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.238490 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8nx2s"] Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.240883 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.244777 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nx2s"] Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.364508 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-utilities\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.364563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-catalog-content\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.364616 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2cwj\" (UniqueName: \"kubernetes.io/projected/952df9e6-2723-40ec-9ad2-34ecd0a9a117-kube-api-access-x2cwj\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.465354 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-utilities\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.465656 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-catalog-content\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.465779 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2cwj\" (UniqueName: \"kubernetes.io/projected/952df9e6-2723-40ec-9ad2-34ecd0a9a117-kube-api-access-x2cwj\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.465914 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-utilities\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.466179 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-catalog-content\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.485071 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2cwj\" (UniqueName: \"kubernetes.io/projected/952df9e6-2723-40ec-9ad2-34ecd0a9a117-kube-api-access-x2cwj\") pod \"redhat-marketplace-8nx2s\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.571621 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:34 crc kubenswrapper[4979]: I1204 12:38:34.987588 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nx2s"] Dec 04 12:38:34 crc kubenswrapper[4979]: W1204 12:38:34.995077 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod952df9e6_2723_40ec_9ad2_34ecd0a9a117.slice/crio-eaad21b5087e4521f47002aa3c072edc4db139e29a48cc0c3a9fc98fdcb09a78 WatchSource:0}: Error finding container eaad21b5087e4521f47002aa3c072edc4db139e29a48cc0c3a9fc98fdcb09a78: Status 404 returned error can't find the container with id eaad21b5087e4521f47002aa3c072edc4db139e29a48cc0c3a9fc98fdcb09a78 Dec 04 12:38:35 crc kubenswrapper[4979]: I1204 12:38:35.099829 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerStarted","Data":"eaad21b5087e4521f47002aa3c072edc4db139e29a48cc0c3a9fc98fdcb09a78"} Dec 04 12:38:35 crc kubenswrapper[4979]: I1204 12:38:35.353007 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:35 crc kubenswrapper[4979]: I1204 12:38:35.353350 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:35 crc kubenswrapper[4979]: I1204 12:38:35.390898 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:36 crc kubenswrapper[4979]: I1204 12:38:36.108851 4979 generic.go:334] "Generic (PLEG): container finished" podID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerID="b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e" exitCode=0 Dec 04 12:38:36 crc kubenswrapper[4979]: I1204 12:38:36.108961 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerDied","Data":"b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e"} Dec 04 12:38:36 crc kubenswrapper[4979]: I1204 12:38:36.168461 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:37 crc kubenswrapper[4979]: I1204 12:38:37.623369 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q4dl2"] Dec 04 12:38:38 crc kubenswrapper[4979]: I1204 12:38:38.126429 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerStarted","Data":"c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6"} Dec 04 12:38:38 crc kubenswrapper[4979]: I1204 12:38:38.126683 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-q4dl2" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="registry-server" containerID="cri-o://5eb32f63726425d8fca1025ea619fc8205e9f9e637480bf5a10a0b9300c4b28b" gracePeriod=2 Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.137507 4979 generic.go:334] "Generic (PLEG): container finished" podID="ce253cac-4481-43be-8091-d15072eaf098" containerID="5eb32f63726425d8fca1025ea619fc8205e9f9e637480bf5a10a0b9300c4b28b" exitCode=0 Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.137552 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerDied","Data":"5eb32f63726425d8fca1025ea619fc8205e9f9e637480bf5a10a0b9300c4b28b"} Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.142512 4979 generic.go:334] "Generic (PLEG): container finished" podID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerID="c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6" exitCode=0 Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.142548 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerDied","Data":"c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6"} Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.687559 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.738631 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-utilities\") pod \"ce253cac-4481-43be-8091-d15072eaf098\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.738756 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g49ht\" (UniqueName: \"kubernetes.io/projected/ce253cac-4481-43be-8091-d15072eaf098-kube-api-access-g49ht\") pod \"ce253cac-4481-43be-8091-d15072eaf098\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.738790 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-catalog-content\") pod \"ce253cac-4481-43be-8091-d15072eaf098\" (UID: \"ce253cac-4481-43be-8091-d15072eaf098\") " Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.739412 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-utilities" (OuterVolumeSpecName: "utilities") pod "ce253cac-4481-43be-8091-d15072eaf098" (UID: "ce253cac-4481-43be-8091-d15072eaf098"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.745068 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce253cac-4481-43be-8091-d15072eaf098-kube-api-access-g49ht" (OuterVolumeSpecName: "kube-api-access-g49ht") pod "ce253cac-4481-43be-8091-d15072eaf098" (UID: "ce253cac-4481-43be-8091-d15072eaf098"). InnerVolumeSpecName "kube-api-access-g49ht". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.841156 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.841207 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g49ht\" (UniqueName: \"kubernetes.io/projected/ce253cac-4481-43be-8091-d15072eaf098-kube-api-access-g49ht\") on node \"crc\" DevicePath \"\"" Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.850690 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce253cac-4481-43be-8091-d15072eaf098" (UID: "ce253cac-4481-43be-8091-d15072eaf098"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:38:39 crc kubenswrapper[4979]: I1204 12:38:39.942633 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce253cac-4481-43be-8091-d15072eaf098-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.153726 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerStarted","Data":"77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae"} Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.156740 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-q4dl2" event={"ID":"ce253cac-4481-43be-8091-d15072eaf098","Type":"ContainerDied","Data":"fef2668763dbe1d5edf1c06eb4a84689aa1db15e8c96bba50707ba3134f468e4"} Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.156810 4979 scope.go:117] "RemoveContainer" containerID="5eb32f63726425d8fca1025ea619fc8205e9f9e637480bf5a10a0b9300c4b28b" Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.156810 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-q4dl2" Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.178341 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8nx2s" podStartSLOduration=2.734410828 podStartE2EDuration="6.178321312s" podCreationTimestamp="2025-12-04 12:38:34 +0000 UTC" firstStartedPulling="2025-12-04 12:38:36.110479395 +0000 UTC m=+3340.384775199" lastFinishedPulling="2025-12-04 12:38:39.554389879 +0000 UTC m=+3343.828685683" observedRunningTime="2025-12-04 12:38:40.172927726 +0000 UTC m=+3344.447223530" watchObservedRunningTime="2025-12-04 12:38:40.178321312 +0000 UTC m=+3344.452617116" Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.180119 4979 scope.go:117] "RemoveContainer" containerID="136c417568ee24e7a878196a8ebcf854e488f0239d74527dfa345a11bcddb3a8" Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.195562 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-q4dl2"] Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.204818 4979 scope.go:117] "RemoveContainer" containerID="ead85f559c6aaed4face0b3fed52dd5281f2c6cf1b7dccafdecaf7ea2885305b" Dec 04 12:38:40 crc kubenswrapper[4979]: I1204 12:38:40.209747 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-q4dl2"] Dec 04 12:38:42 crc kubenswrapper[4979]: I1204 12:38:42.207550 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce253cac-4481-43be-8091-d15072eaf098" path="/var/lib/kubelet/pods/ce253cac-4481-43be-8091-d15072eaf098/volumes" Dec 04 12:38:44 crc kubenswrapper[4979]: I1204 12:38:44.572513 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:44 crc kubenswrapper[4979]: I1204 12:38:44.572934 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:44 crc kubenswrapper[4979]: I1204 12:38:44.614076 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:45 crc kubenswrapper[4979]: I1204 12:38:45.453551 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:46 crc kubenswrapper[4979]: I1204 12:38:46.021197 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nx2s"] Dec 04 12:38:47 crc kubenswrapper[4979]: I1204 12:38:47.429850 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8nx2s" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="registry-server" containerID="cri-o://77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae" gracePeriod=2 Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.420509 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.438057 4979 generic.go:334] "Generic (PLEG): container finished" podID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerID="77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae" exitCode=0 Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.438106 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerDied","Data":"77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae"} Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.438135 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8nx2s" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.438155 4979 scope.go:117] "RemoveContainer" containerID="77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.438141 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8nx2s" event={"ID":"952df9e6-2723-40ec-9ad2-34ecd0a9a117","Type":"ContainerDied","Data":"eaad21b5087e4521f47002aa3c072edc4db139e29a48cc0c3a9fc98fdcb09a78"} Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.463418 4979 scope.go:117] "RemoveContainer" containerID="c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.492827 4979 scope.go:117] "RemoveContainer" containerID="b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.512093 4979 scope.go:117] "RemoveContainer" containerID="77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae" Dec 04 12:38:48 crc kubenswrapper[4979]: E1204 12:38:48.512512 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae\": container with ID starting with 77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae not found: ID does not exist" containerID="77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.512541 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae"} err="failed to get container status \"77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae\": rpc error: code = NotFound desc = could not find container \"77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae\": container with ID starting with 77d4b6608c6a10d6e613ec49151e2cca3567ea40711bd0f33eb77fe301cc2bae not found: ID does not exist" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.512563 4979 scope.go:117] "RemoveContainer" containerID="c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6" Dec 04 12:38:48 crc kubenswrapper[4979]: E1204 12:38:48.512909 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6\": container with ID starting with c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6 not found: ID does not exist" containerID="c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.512961 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6"} err="failed to get container status \"c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6\": rpc error: code = NotFound desc = could not find container \"c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6\": container with ID starting with c0c81580d0f395131491f58f7d136e171b0480cb2bd4925c05ca9c03a854afc6 not found: ID does not exist" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.512994 4979 scope.go:117] "RemoveContainer" containerID="b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e" Dec 04 12:38:48 crc kubenswrapper[4979]: E1204 12:38:48.513572 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e\": container with ID starting with b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e not found: ID does not exist" containerID="b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.513597 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e"} err="failed to get container status \"b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e\": rpc error: code = NotFound desc = could not find container \"b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e\": container with ID starting with b4a76af86a46c5c1b4ac3cfe3df05453b7ddd808e62893f239f79be108d4db5e not found: ID does not exist" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.551553 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2cwj\" (UniqueName: \"kubernetes.io/projected/952df9e6-2723-40ec-9ad2-34ecd0a9a117-kube-api-access-x2cwj\") pod \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.551701 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-catalog-content\") pod \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.551743 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-utilities\") pod \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\" (UID: \"952df9e6-2723-40ec-9ad2-34ecd0a9a117\") " Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.552630 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-utilities" (OuterVolumeSpecName: "utilities") pod "952df9e6-2723-40ec-9ad2-34ecd0a9a117" (UID: "952df9e6-2723-40ec-9ad2-34ecd0a9a117"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.557094 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/952df9e6-2723-40ec-9ad2-34ecd0a9a117-kube-api-access-x2cwj" (OuterVolumeSpecName: "kube-api-access-x2cwj") pod "952df9e6-2723-40ec-9ad2-34ecd0a9a117" (UID: "952df9e6-2723-40ec-9ad2-34ecd0a9a117"). InnerVolumeSpecName "kube-api-access-x2cwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.569014 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "952df9e6-2723-40ec-9ad2-34ecd0a9a117" (UID: "952df9e6-2723-40ec-9ad2-34ecd0a9a117"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.652848 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.652887 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2cwj\" (UniqueName: \"kubernetes.io/projected/952df9e6-2723-40ec-9ad2-34ecd0a9a117-kube-api-access-x2cwj\") on node \"crc\" DevicePath \"\"" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.652898 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/952df9e6-2723-40ec-9ad2-34ecd0a9a117-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.767198 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nx2s"] Dec 04 12:38:48 crc kubenswrapper[4979]: I1204 12:38:48.772278 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8nx2s"] Dec 04 12:38:50 crc kubenswrapper[4979]: I1204 12:38:50.218920 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" path="/var/lib/kubelet/pods/952df9e6-2723-40ec-9ad2-34ecd0a9a117/volumes" Dec 04 12:38:58 crc kubenswrapper[4979]: I1204 12:38:58.040449 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:38:58 crc kubenswrapper[4979]: I1204 12:38:58.040877 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:39:28 crc kubenswrapper[4979]: I1204 12:39:28.041247 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:39:28 crc kubenswrapper[4979]: I1204 12:39:28.041930 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.040447 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.041398 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.041597 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.042822 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.042935 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f" gracePeriod=600 Dec 04 12:39:58 crc kubenswrapper[4979]: E1204 12:39:58.064851 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9b48d97_5715_467e_b044_4812d6dfea63.slice/crio-da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f.scope\": RecentStats: unable to find data in memory cache]" Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.972320 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f" exitCode=0 Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.972334 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f"} Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.972725 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0"} Dec 04 12:39:58 crc kubenswrapper[4979]: I1204 12:39:58.972758 4979 scope.go:117] "RemoveContainer" containerID="387b5df76a26059ebadff96bec17d0fa8baf833b542077edd858f0d4a77d111d" Dec 04 12:41:58 crc kubenswrapper[4979]: I1204 12:41:58.040822 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:41:58 crc kubenswrapper[4979]: I1204 12:41:58.041270 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:42:28 crc kubenswrapper[4979]: I1204 12:42:28.041244 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:42:28 crc kubenswrapper[4979]: I1204 12:42:28.041884 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.040762 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.041429 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.041482 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.042156 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.042217 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" gracePeriod=600 Dec 04 12:42:58 crc kubenswrapper[4979]: E1204 12:42:58.161854 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.242757 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" exitCode=0 Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.242820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0"} Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.242891 4979 scope.go:117] "RemoveContainer" containerID="da359468d56b7b11d8d0b358f5f6c1c313ce72b88047378f84677312b5f2aa4f" Dec 04 12:42:58 crc kubenswrapper[4979]: I1204 12:42:58.243399 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:42:58 crc kubenswrapper[4979]: E1204 12:42:58.243678 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:43:09 crc kubenswrapper[4979]: I1204 12:43:09.199654 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:43:09 crc kubenswrapper[4979]: E1204 12:43:09.200410 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:43:22 crc kubenswrapper[4979]: I1204 12:43:22.198654 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:43:22 crc kubenswrapper[4979]: E1204 12:43:22.199398 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:43:35 crc kubenswrapper[4979]: I1204 12:43:35.199224 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:43:35 crc kubenswrapper[4979]: E1204 12:43:35.200137 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:43:50 crc kubenswrapper[4979]: I1204 12:43:50.199362 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:43:50 crc kubenswrapper[4979]: E1204 12:43:50.199971 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:44:01 crc kubenswrapper[4979]: I1204 12:44:01.200217 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:44:01 crc kubenswrapper[4979]: E1204 12:44:01.201356 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:44:16 crc kubenswrapper[4979]: I1204 12:44:16.204334 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:44:16 crc kubenswrapper[4979]: E1204 12:44:16.205241 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.776804 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9hq4k"] Dec 04 12:44:18 crc kubenswrapper[4979]: E1204 12:44:18.777099 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="registry-server" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777111 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="registry-server" Dec 04 12:44:18 crc kubenswrapper[4979]: E1204 12:44:18.777127 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="extract-utilities" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777134 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="extract-utilities" Dec 04 12:44:18 crc kubenswrapper[4979]: E1204 12:44:18.777146 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="extract-content" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777152 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="extract-content" Dec 04 12:44:18 crc kubenswrapper[4979]: E1204 12:44:18.777164 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="extract-content" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777170 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="extract-content" Dec 04 12:44:18 crc kubenswrapper[4979]: E1204 12:44:18.777180 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="extract-utilities" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777185 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="extract-utilities" Dec 04 12:44:18 crc kubenswrapper[4979]: E1204 12:44:18.777205 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="registry-server" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777210 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="registry-server" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777376 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="952df9e6-2723-40ec-9ad2-34ecd0a9a117" containerName="registry-server" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.777395 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce253cac-4481-43be-8091-d15072eaf098" containerName="registry-server" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.778428 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.789187 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9hq4k"] Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.967533 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-catalog-content\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.967584 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qc5j\" (UniqueName: \"kubernetes.io/projected/e70885e7-6962-4a29-a271-a97b148342ba-kube-api-access-6qc5j\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:18 crc kubenswrapper[4979]: I1204 12:44:18.967610 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-utilities\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.069391 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-catalog-content\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.069472 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qc5j\" (UniqueName: \"kubernetes.io/projected/e70885e7-6962-4a29-a271-a97b148342ba-kube-api-access-6qc5j\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.069510 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-utilities\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.070433 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-catalog-content\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.070500 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-utilities\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.096716 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qc5j\" (UniqueName: \"kubernetes.io/projected/e70885e7-6962-4a29-a271-a97b148342ba-kube-api-access-6qc5j\") pod \"certified-operators-9hq4k\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.098630 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.377884 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9hq4k"] Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.865251 4979 generic.go:334] "Generic (PLEG): container finished" podID="e70885e7-6962-4a29-a271-a97b148342ba" containerID="de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398" exitCode=0 Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.865354 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hq4k" event={"ID":"e70885e7-6962-4a29-a271-a97b148342ba","Type":"ContainerDied","Data":"de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398"} Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.866404 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hq4k" event={"ID":"e70885e7-6962-4a29-a271-a97b148342ba","Type":"ContainerStarted","Data":"28892ee842be8ddff490ea7eb8174be5a41f3d7db9180efe926530051a1c5e48"} Dec 04 12:44:19 crc kubenswrapper[4979]: I1204 12:44:19.867006 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:44:21 crc kubenswrapper[4979]: I1204 12:44:21.897170 4979 generic.go:334] "Generic (PLEG): container finished" podID="e70885e7-6962-4a29-a271-a97b148342ba" containerID="0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf" exitCode=0 Dec 04 12:44:21 crc kubenswrapper[4979]: I1204 12:44:21.897267 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hq4k" event={"ID":"e70885e7-6962-4a29-a271-a97b148342ba","Type":"ContainerDied","Data":"0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf"} Dec 04 12:44:22 crc kubenswrapper[4979]: I1204 12:44:22.906503 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hq4k" event={"ID":"e70885e7-6962-4a29-a271-a97b148342ba","Type":"ContainerStarted","Data":"8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3"} Dec 04 12:44:22 crc kubenswrapper[4979]: I1204 12:44:22.928638 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9hq4k" podStartSLOduration=2.447112727 podStartE2EDuration="4.928614446s" podCreationTimestamp="2025-12-04 12:44:18 +0000 UTC" firstStartedPulling="2025-12-04 12:44:19.866735308 +0000 UTC m=+3684.141031122" lastFinishedPulling="2025-12-04 12:44:22.348237037 +0000 UTC m=+3686.622532841" observedRunningTime="2025-12-04 12:44:22.923769754 +0000 UTC m=+3687.198065568" watchObservedRunningTime="2025-12-04 12:44:22.928614446 +0000 UTC m=+3687.202910250" Dec 04 12:44:29 crc kubenswrapper[4979]: I1204 12:44:29.099240 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:29 crc kubenswrapper[4979]: I1204 12:44:29.100044 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:29 crc kubenswrapper[4979]: I1204 12:44:29.143292 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:29 crc kubenswrapper[4979]: I1204 12:44:29.991203 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:30 crc kubenswrapper[4979]: I1204 12:44:30.031091 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9hq4k"] Dec 04 12:44:31 crc kubenswrapper[4979]: I1204 12:44:31.198892 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:44:31 crc kubenswrapper[4979]: E1204 12:44:31.199110 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:44:31 crc kubenswrapper[4979]: I1204 12:44:31.967570 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9hq4k" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="registry-server" containerID="cri-o://8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3" gracePeriod=2 Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.433823 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.574760 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-utilities\") pod \"e70885e7-6962-4a29-a271-a97b148342ba\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.574871 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qc5j\" (UniqueName: \"kubernetes.io/projected/e70885e7-6962-4a29-a271-a97b148342ba-kube-api-access-6qc5j\") pod \"e70885e7-6962-4a29-a271-a97b148342ba\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.575047 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-catalog-content\") pod \"e70885e7-6962-4a29-a271-a97b148342ba\" (UID: \"e70885e7-6962-4a29-a271-a97b148342ba\") " Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.575769 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-utilities" (OuterVolumeSpecName: "utilities") pod "e70885e7-6962-4a29-a271-a97b148342ba" (UID: "e70885e7-6962-4a29-a271-a97b148342ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.576515 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.580812 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e70885e7-6962-4a29-a271-a97b148342ba-kube-api-access-6qc5j" (OuterVolumeSpecName: "kube-api-access-6qc5j") pod "e70885e7-6962-4a29-a271-a97b148342ba" (UID: "e70885e7-6962-4a29-a271-a97b148342ba"). InnerVolumeSpecName "kube-api-access-6qc5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.621474 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e70885e7-6962-4a29-a271-a97b148342ba" (UID: "e70885e7-6962-4a29-a271-a97b148342ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.677933 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e70885e7-6962-4a29-a271-a97b148342ba-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.677974 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qc5j\" (UniqueName: \"kubernetes.io/projected/e70885e7-6962-4a29-a271-a97b148342ba-kube-api-access-6qc5j\") on node \"crc\" DevicePath \"\"" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.983645 4979 generic.go:334] "Generic (PLEG): container finished" podID="e70885e7-6962-4a29-a271-a97b148342ba" containerID="8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3" exitCode=0 Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.983691 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hq4k" event={"ID":"e70885e7-6962-4a29-a271-a97b148342ba","Type":"ContainerDied","Data":"8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3"} Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.983788 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9hq4k" event={"ID":"e70885e7-6962-4a29-a271-a97b148342ba","Type":"ContainerDied","Data":"28892ee842be8ddff490ea7eb8174be5a41f3d7db9180efe926530051a1c5e48"} Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.983821 4979 scope.go:117] "RemoveContainer" containerID="8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3" Dec 04 12:44:33 crc kubenswrapper[4979]: I1204 12:44:33.983876 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9hq4k" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:33.999704 4979 scope.go:117] "RemoveContainer" containerID="0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.018156 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9hq4k"] Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.026234 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9hq4k"] Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.026535 4979 scope.go:117] "RemoveContainer" containerID="de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.045666 4979 scope.go:117] "RemoveContainer" containerID="8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3" Dec 04 12:44:34 crc kubenswrapper[4979]: E1204 12:44:34.046372 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3\": container with ID starting with 8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3 not found: ID does not exist" containerID="8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.046434 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3"} err="failed to get container status \"8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3\": rpc error: code = NotFound desc = could not find container \"8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3\": container with ID starting with 8616d163f502483551be31d62e5fd3e4d8ceab70b7af94b6f9fe7f47d99c22a3 not found: ID does not exist" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.046468 4979 scope.go:117] "RemoveContainer" containerID="0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf" Dec 04 12:44:34 crc kubenswrapper[4979]: E1204 12:44:34.046911 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf\": container with ID starting with 0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf not found: ID does not exist" containerID="0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.046940 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf"} err="failed to get container status \"0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf\": rpc error: code = NotFound desc = could not find container \"0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf\": container with ID starting with 0b8d9623ce929a6f7603405c718137679edf7ce0e5c2065c7916d7397af1fccf not found: ID does not exist" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.046960 4979 scope.go:117] "RemoveContainer" containerID="de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398" Dec 04 12:44:34 crc kubenswrapper[4979]: E1204 12:44:34.047257 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398\": container with ID starting with de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398 not found: ID does not exist" containerID="de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.047290 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398"} err="failed to get container status \"de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398\": rpc error: code = NotFound desc = could not find container \"de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398\": container with ID starting with de1fc253f0fe764d7d4c034f2085a2be320e1e17bed4b1cf7b5d0665ec9b3398 not found: ID does not exist" Dec 04 12:44:34 crc kubenswrapper[4979]: I1204 12:44:34.209885 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e70885e7-6962-4a29-a271-a97b148342ba" path="/var/lib/kubelet/pods/e70885e7-6962-4a29-a271-a97b148342ba/volumes" Dec 04 12:44:46 crc kubenswrapper[4979]: I1204 12:44:46.202120 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:44:46 crc kubenswrapper[4979]: E1204 12:44:46.202794 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:44:58 crc kubenswrapper[4979]: I1204 12:44:58.199171 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:44:58 crc kubenswrapper[4979]: E1204 12:44:58.199889 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.156400 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b"] Dec 04 12:45:00 crc kubenswrapper[4979]: E1204 12:45:00.156755 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="extract-utilities" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.156771 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="extract-utilities" Dec 04 12:45:00 crc kubenswrapper[4979]: E1204 12:45:00.156843 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="registry-server" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.156850 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="registry-server" Dec 04 12:45:00 crc kubenswrapper[4979]: E1204 12:45:00.156868 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="extract-content" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.156876 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="extract-content" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.157075 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e70885e7-6962-4a29-a271-a97b148342ba" containerName="registry-server" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.157647 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.160759 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.161105 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.174376 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b"] Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.335862 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0551bbc-e978-423a-8210-1f99102787a7-secret-volume\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.336529 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pd4jd\" (UniqueName: \"kubernetes.io/projected/d0551bbc-e978-423a-8210-1f99102787a7-kube-api-access-pd4jd\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.336712 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0551bbc-e978-423a-8210-1f99102787a7-config-volume\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.438021 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0551bbc-e978-423a-8210-1f99102787a7-config-volume\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.438109 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0551bbc-e978-423a-8210-1f99102787a7-secret-volume\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.438171 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pd4jd\" (UniqueName: \"kubernetes.io/projected/d0551bbc-e978-423a-8210-1f99102787a7-kube-api-access-pd4jd\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.439195 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0551bbc-e978-423a-8210-1f99102787a7-config-volume\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.444828 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0551bbc-e978-423a-8210-1f99102787a7-secret-volume\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.458994 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pd4jd\" (UniqueName: \"kubernetes.io/projected/d0551bbc-e978-423a-8210-1f99102787a7-kube-api-access-pd4jd\") pod \"collect-profiles-29414205-m4m2b\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.476751 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:00 crc kubenswrapper[4979]: I1204 12:45:00.905417 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b"] Dec 04 12:45:01 crc kubenswrapper[4979]: I1204 12:45:01.183994 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" event={"ID":"d0551bbc-e978-423a-8210-1f99102787a7","Type":"ContainerStarted","Data":"de9062458b15276a25c92d4889ff2351a826d3665770cbe19b7918dac8f98a8e"} Dec 04 12:45:01 crc kubenswrapper[4979]: I1204 12:45:01.184390 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" event={"ID":"d0551bbc-e978-423a-8210-1f99102787a7","Type":"ContainerStarted","Data":"2de4173a05769ecf513d7d2f465ebfb7d5b5c94f50a13dfccd5ea8ec139ccb33"} Dec 04 12:45:01 crc kubenswrapper[4979]: I1204 12:45:01.207817 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" podStartSLOduration=1.207798781 podStartE2EDuration="1.207798781s" podCreationTimestamp="2025-12-04 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 12:45:01.200714048 +0000 UTC m=+3725.475009852" watchObservedRunningTime="2025-12-04 12:45:01.207798781 +0000 UTC m=+3725.482094585" Dec 04 12:45:02 crc kubenswrapper[4979]: I1204 12:45:02.192735 4979 generic.go:334] "Generic (PLEG): container finished" podID="d0551bbc-e978-423a-8210-1f99102787a7" containerID="de9062458b15276a25c92d4889ff2351a826d3665770cbe19b7918dac8f98a8e" exitCode=0 Dec 04 12:45:02 crc kubenswrapper[4979]: I1204 12:45:02.192779 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" event={"ID":"d0551bbc-e978-423a-8210-1f99102787a7","Type":"ContainerDied","Data":"de9062458b15276a25c92d4889ff2351a826d3665770cbe19b7918dac8f98a8e"} Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.468021 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.585544 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0551bbc-e978-423a-8210-1f99102787a7-config-volume\") pod \"d0551bbc-e978-423a-8210-1f99102787a7\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.585700 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pd4jd\" (UniqueName: \"kubernetes.io/projected/d0551bbc-e978-423a-8210-1f99102787a7-kube-api-access-pd4jd\") pod \"d0551bbc-e978-423a-8210-1f99102787a7\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.585734 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0551bbc-e978-423a-8210-1f99102787a7-secret-volume\") pod \"d0551bbc-e978-423a-8210-1f99102787a7\" (UID: \"d0551bbc-e978-423a-8210-1f99102787a7\") " Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.587050 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0551bbc-e978-423a-8210-1f99102787a7-config-volume" (OuterVolumeSpecName: "config-volume") pod "d0551bbc-e978-423a-8210-1f99102787a7" (UID: "d0551bbc-e978-423a-8210-1f99102787a7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.590872 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0551bbc-e978-423a-8210-1f99102787a7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d0551bbc-e978-423a-8210-1f99102787a7" (UID: "d0551bbc-e978-423a-8210-1f99102787a7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.591460 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0551bbc-e978-423a-8210-1f99102787a7-kube-api-access-pd4jd" (OuterVolumeSpecName: "kube-api-access-pd4jd") pod "d0551bbc-e978-423a-8210-1f99102787a7" (UID: "d0551bbc-e978-423a-8210-1f99102787a7"). InnerVolumeSpecName "kube-api-access-pd4jd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.688707 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pd4jd\" (UniqueName: \"kubernetes.io/projected/d0551bbc-e978-423a-8210-1f99102787a7-kube-api-access-pd4jd\") on node \"crc\" DevicePath \"\"" Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.689092 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d0551bbc-e978-423a-8210-1f99102787a7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:45:03 crc kubenswrapper[4979]: I1204 12:45:03.689108 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d0551bbc-e978-423a-8210-1f99102787a7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 12:45:04 crc kubenswrapper[4979]: I1204 12:45:04.207373 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" event={"ID":"d0551bbc-e978-423a-8210-1f99102787a7","Type":"ContainerDied","Data":"2de4173a05769ecf513d7d2f465ebfb7d5b5c94f50a13dfccd5ea8ec139ccb33"} Dec 04 12:45:04 crc kubenswrapper[4979]: I1204 12:45:04.207421 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2de4173a05769ecf513d7d2f465ebfb7d5b5c94f50a13dfccd5ea8ec139ccb33" Dec 04 12:45:04 crc kubenswrapper[4979]: I1204 12:45:04.207438 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b" Dec 04 12:45:04 crc kubenswrapper[4979]: I1204 12:45:04.271143 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p"] Dec 04 12:45:04 crc kubenswrapper[4979]: I1204 12:45:04.277456 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414160-q7t8p"] Dec 04 12:45:06 crc kubenswrapper[4979]: I1204 12:45:06.212607 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a94915fc-616b-4eb7-9133-19411e6ab093" path="/var/lib/kubelet/pods/a94915fc-616b-4eb7-9133-19411e6ab093/volumes" Dec 04 12:45:09 crc kubenswrapper[4979]: I1204 12:45:09.199478 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:45:09 crc kubenswrapper[4979]: E1204 12:45:09.199713 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:45:23 crc kubenswrapper[4979]: I1204 12:45:23.198921 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:45:23 crc kubenswrapper[4979]: E1204 12:45:23.199639 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:45:37 crc kubenswrapper[4979]: I1204 12:45:37.199850 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:45:37 crc kubenswrapper[4979]: E1204 12:45:37.200670 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:45:49 crc kubenswrapper[4979]: I1204 12:45:49.199766 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:45:49 crc kubenswrapper[4979]: E1204 12:45:49.201162 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:46:00 crc kubenswrapper[4979]: I1204 12:46:00.199045 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:46:00 crc kubenswrapper[4979]: E1204 12:46:00.199930 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:46:05 crc kubenswrapper[4979]: I1204 12:46:05.380726 4979 scope.go:117] "RemoveContainer" containerID="02caaddbcb1c5158efe7f727897e2e968d74aa9a753cf1f03267474ede0c3f89" Dec 04 12:46:11 crc kubenswrapper[4979]: I1204 12:46:11.199484 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:46:11 crc kubenswrapper[4979]: E1204 12:46:11.200551 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:46:24 crc kubenswrapper[4979]: I1204 12:46:24.199930 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:46:24 crc kubenswrapper[4979]: E1204 12:46:24.200808 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:46:37 crc kubenswrapper[4979]: I1204 12:46:37.199451 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:46:37 crc kubenswrapper[4979]: E1204 12:46:37.200474 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:46:49 crc kubenswrapper[4979]: I1204 12:46:49.198957 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:46:49 crc kubenswrapper[4979]: E1204 12:46:49.199835 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:47:02 crc kubenswrapper[4979]: I1204 12:47:02.199476 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:47:02 crc kubenswrapper[4979]: E1204 12:47:02.200563 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:47:16 crc kubenswrapper[4979]: I1204 12:47:16.199649 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:47:16 crc kubenswrapper[4979]: E1204 12:47:16.201008 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:47:28 crc kubenswrapper[4979]: I1204 12:47:28.198756 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:47:28 crc kubenswrapper[4979]: E1204 12:47:28.199595 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.541133 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-w98xz"] Dec 04 12:47:41 crc kubenswrapper[4979]: E1204 12:47:41.542036 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0551bbc-e978-423a-8210-1f99102787a7" containerName="collect-profiles" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.542052 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0551bbc-e978-423a-8210-1f99102787a7" containerName="collect-profiles" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.542245 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0551bbc-e978-423a-8210-1f99102787a7" containerName="collect-profiles" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.543415 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.556068 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w98xz"] Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.729804 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-utilities\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.729914 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-catalog-content\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.729960 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmvqb\" (UniqueName: \"kubernetes.io/projected/74f9fae7-8a16-48da-a70a-153bdf67800f-kube-api-access-hmvqb\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.831219 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-catalog-content\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.831286 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmvqb\" (UniqueName: \"kubernetes.io/projected/74f9fae7-8a16-48da-a70a-153bdf67800f-kube-api-access-hmvqb\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.831365 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-utilities\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.831901 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-catalog-content\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:41 crc kubenswrapper[4979]: I1204 12:47:41.831963 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-utilities\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:42 crc kubenswrapper[4979]: I1204 12:47:42.071138 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmvqb\" (UniqueName: \"kubernetes.io/projected/74f9fae7-8a16-48da-a70a-153bdf67800f-kube-api-access-hmvqb\") pod \"community-operators-w98xz\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:42 crc kubenswrapper[4979]: I1204 12:47:42.173533 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:42 crc kubenswrapper[4979]: I1204 12:47:42.198803 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:47:42 crc kubenswrapper[4979]: E1204 12:47:42.199099 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:47:42 crc kubenswrapper[4979]: I1204 12:47:42.396649 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-w98xz"] Dec 04 12:47:42 crc kubenswrapper[4979]: I1204 12:47:42.442201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w98xz" event={"ID":"74f9fae7-8a16-48da-a70a-153bdf67800f","Type":"ContainerStarted","Data":"c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944"} Dec 04 12:47:43 crc kubenswrapper[4979]: I1204 12:47:43.451115 4979 generic.go:334] "Generic (PLEG): container finished" podID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerID="f598840c7b39dd8510d8bebf2e7f1466333095766a1a2f1e1fe97b7a35663573" exitCode=0 Dec 04 12:47:43 crc kubenswrapper[4979]: I1204 12:47:43.451186 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w98xz" event={"ID":"74f9fae7-8a16-48da-a70a-153bdf67800f","Type":"ContainerDied","Data":"f598840c7b39dd8510d8bebf2e7f1466333095766a1a2f1e1fe97b7a35663573"} Dec 04 12:47:44 crc kubenswrapper[4979]: I1204 12:47:44.461688 4979 generic.go:334] "Generic (PLEG): container finished" podID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerID="07a0cc8e8e099ba24746bd9ba4f9704135c0c0272ed10969f490496d1aaccca6" exitCode=0 Dec 04 12:47:44 crc kubenswrapper[4979]: I1204 12:47:44.461739 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w98xz" event={"ID":"74f9fae7-8a16-48da-a70a-153bdf67800f","Type":"ContainerDied","Data":"07a0cc8e8e099ba24746bd9ba4f9704135c0c0272ed10969f490496d1aaccca6"} Dec 04 12:47:45 crc kubenswrapper[4979]: I1204 12:47:45.470582 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w98xz" event={"ID":"74f9fae7-8a16-48da-a70a-153bdf67800f","Type":"ContainerStarted","Data":"acf4bd1a70c05221d03b52a78eff41d891cc17b7172f4f091967d5ab56bd9698"} Dec 04 12:47:45 crc kubenswrapper[4979]: I1204 12:47:45.497763 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-w98xz" podStartSLOduration=3.100738888 podStartE2EDuration="4.497739733s" podCreationTimestamp="2025-12-04 12:47:41 +0000 UTC" firstStartedPulling="2025-12-04 12:47:43.452632046 +0000 UTC m=+3887.726927840" lastFinishedPulling="2025-12-04 12:47:44.849632881 +0000 UTC m=+3889.123928685" observedRunningTime="2025-12-04 12:47:45.491204165 +0000 UTC m=+3889.765500019" watchObservedRunningTime="2025-12-04 12:47:45.497739733 +0000 UTC m=+3889.772035557" Dec 04 12:47:52 crc kubenswrapper[4979]: I1204 12:47:52.174131 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:52 crc kubenswrapper[4979]: I1204 12:47:52.174982 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:52 crc kubenswrapper[4979]: I1204 12:47:52.218403 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:52 crc kubenswrapper[4979]: I1204 12:47:52.704480 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:52 crc kubenswrapper[4979]: I1204 12:47:52.760717 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w98xz"] Dec 04 12:47:54 crc kubenswrapper[4979]: I1204 12:47:54.539918 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-w98xz" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="registry-server" containerID="cri-o://acf4bd1a70c05221d03b52a78eff41d891cc17b7172f4f091967d5ab56bd9698" gracePeriod=2 Dec 04 12:47:55 crc kubenswrapper[4979]: I1204 12:47:55.548870 4979 generic.go:334] "Generic (PLEG): container finished" podID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerID="acf4bd1a70c05221d03b52a78eff41d891cc17b7172f4f091967d5ab56bd9698" exitCode=0 Dec 04 12:47:55 crc kubenswrapper[4979]: I1204 12:47:55.548943 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w98xz" event={"ID":"74f9fae7-8a16-48da-a70a-153bdf67800f","Type":"ContainerDied","Data":"acf4bd1a70c05221d03b52a78eff41d891cc17b7172f4f091967d5ab56bd9698"} Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.033828 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.135915 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hmvqb\" (UniqueName: \"kubernetes.io/projected/74f9fae7-8a16-48da-a70a-153bdf67800f-kube-api-access-hmvqb\") pod \"74f9fae7-8a16-48da-a70a-153bdf67800f\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.136050 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-catalog-content\") pod \"74f9fae7-8a16-48da-a70a-153bdf67800f\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.136086 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-utilities\") pod \"74f9fae7-8a16-48da-a70a-153bdf67800f\" (UID: \"74f9fae7-8a16-48da-a70a-153bdf67800f\") " Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.137772 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-utilities" (OuterVolumeSpecName: "utilities") pod "74f9fae7-8a16-48da-a70a-153bdf67800f" (UID: "74f9fae7-8a16-48da-a70a-153bdf67800f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.143785 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74f9fae7-8a16-48da-a70a-153bdf67800f-kube-api-access-hmvqb" (OuterVolumeSpecName: "kube-api-access-hmvqb") pod "74f9fae7-8a16-48da-a70a-153bdf67800f" (UID: "74f9fae7-8a16-48da-a70a-153bdf67800f"). InnerVolumeSpecName "kube-api-access-hmvqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.193136 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74f9fae7-8a16-48da-a70a-153bdf67800f" (UID: "74f9fae7-8a16-48da-a70a-153bdf67800f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.204414 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:47:56 crc kubenswrapper[4979]: E1204 12:47:56.204600 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.237482 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.237534 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74f9fae7-8a16-48da-a70a-153bdf67800f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.237549 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hmvqb\" (UniqueName: \"kubernetes.io/projected/74f9fae7-8a16-48da-a70a-153bdf67800f-kube-api-access-hmvqb\") on node \"crc\" DevicePath \"\"" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.557436 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-w98xz" event={"ID":"74f9fae7-8a16-48da-a70a-153bdf67800f","Type":"ContainerDied","Data":"c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944"} Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.557522 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-w98xz" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.557835 4979 scope.go:117] "RemoveContainer" containerID="acf4bd1a70c05221d03b52a78eff41d891cc17b7172f4f091967d5ab56bd9698" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.576260 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-w98xz"] Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.578057 4979 scope.go:117] "RemoveContainer" containerID="07a0cc8e8e099ba24746bd9ba4f9704135c0c0272ed10969f490496d1aaccca6" Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.582873 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-w98xz"] Dec 04 12:47:56 crc kubenswrapper[4979]: I1204 12:47:56.598847 4979 scope.go:117] "RemoveContainer" containerID="f598840c7b39dd8510d8bebf2e7f1466333095766a1a2f1e1fe97b7a35663573" Dec 04 12:47:56 crc kubenswrapper[4979]: E1204 12:47:56.803477 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice/crio-c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice\": RecentStats: unable to find data in memory cache]" Dec 04 12:47:58 crc kubenswrapper[4979]: I1204 12:47:58.211244 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" path="/var/lib/kubelet/pods/74f9fae7-8a16-48da-a70a-153bdf67800f/volumes" Dec 04 12:48:06 crc kubenswrapper[4979]: E1204 12:48:06.971630 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice/crio-c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944\": RecentStats: unable to find data in memory cache]" Dec 04 12:48:11 crc kubenswrapper[4979]: I1204 12:48:11.198732 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:48:12 crc kubenswrapper[4979]: I1204 12:48:12.678786 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"3c227c4073a3b2038c807150c1376009b949d75a319f0bf4a7c3711918aa7c2d"} Dec 04 12:48:17 crc kubenswrapper[4979]: E1204 12:48:17.180959 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice/crio-c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice\": RecentStats: unable to find data in memory cache]" Dec 04 12:48:27 crc kubenswrapper[4979]: E1204 12:48:27.363262 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice/crio-c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944\": RecentStats: unable to find data in memory cache]" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.767535 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nq9bk"] Dec 04 12:48:27 crc kubenswrapper[4979]: E1204 12:48:27.768388 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="registry-server" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.768410 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="registry-server" Dec 04 12:48:27 crc kubenswrapper[4979]: E1204 12:48:27.768441 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="extract-content" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.768450 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="extract-content" Dec 04 12:48:27 crc kubenswrapper[4979]: E1204 12:48:27.768476 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="extract-utilities" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.768486 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="extract-utilities" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.768678 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f9fae7-8a16-48da-a70a-153bdf67800f" containerName="registry-server" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.769967 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.788191 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nq9bk"] Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.882662 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-utilities\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.882709 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwbkd\" (UniqueName: \"kubernetes.io/projected/b752c1f8-051c-43f7-a762-080e606c76e8-kube-api-access-gwbkd\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.882733 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-catalog-content\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.985146 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-utilities\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.985262 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwbkd\" (UniqueName: \"kubernetes.io/projected/b752c1f8-051c-43f7-a762-080e606c76e8-kube-api-access-gwbkd\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.985750 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-utilities\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.986069 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-catalog-content\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:27 crc kubenswrapper[4979]: I1204 12:48:27.986180 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-catalog-content\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:28 crc kubenswrapper[4979]: I1204 12:48:28.010316 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwbkd\" (UniqueName: \"kubernetes.io/projected/b752c1f8-051c-43f7-a762-080e606c76e8-kube-api-access-gwbkd\") pod \"redhat-operators-nq9bk\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:28 crc kubenswrapper[4979]: I1204 12:48:28.096348 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:28 crc kubenswrapper[4979]: I1204 12:48:28.550143 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nq9bk"] Dec 04 12:48:28 crc kubenswrapper[4979]: I1204 12:48:28.824778 4979 generic.go:334] "Generic (PLEG): container finished" podID="b752c1f8-051c-43f7-a762-080e606c76e8" containerID="3e90ecf475d711b1a15c8dcd8f9e75e939e5e51e617948222d582489f8967e8f" exitCode=0 Dec 04 12:48:28 crc kubenswrapper[4979]: I1204 12:48:28.824832 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerDied","Data":"3e90ecf475d711b1a15c8dcd8f9e75e939e5e51e617948222d582489f8967e8f"} Dec 04 12:48:28 crc kubenswrapper[4979]: I1204 12:48:28.824865 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerStarted","Data":"19b7deae8906b2a68886179f25b7c7ee6a2036ce5754f8664e142c786b671d0a"} Dec 04 12:48:30 crc kubenswrapper[4979]: I1204 12:48:30.839882 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerStarted","Data":"35fcaa73e19ba34fd5e867f5a7a7bfc18da9e41d5f63d38321cb0b8b0c3d3257"} Dec 04 12:48:31 crc kubenswrapper[4979]: I1204 12:48:31.848188 4979 generic.go:334] "Generic (PLEG): container finished" podID="b752c1f8-051c-43f7-a762-080e606c76e8" containerID="35fcaa73e19ba34fd5e867f5a7a7bfc18da9e41d5f63d38321cb0b8b0c3d3257" exitCode=0 Dec 04 12:48:31 crc kubenswrapper[4979]: I1204 12:48:31.848281 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerDied","Data":"35fcaa73e19ba34fd5e867f5a7a7bfc18da9e41d5f63d38321cb0b8b0c3d3257"} Dec 04 12:48:32 crc kubenswrapper[4979]: I1204 12:48:32.857843 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerStarted","Data":"ce543bc29510b3c1595cf619094593bcb99dc7ce107829a0676ba987228a00f6"} Dec 04 12:48:32 crc kubenswrapper[4979]: I1204 12:48:32.880457 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nq9bk" podStartSLOduration=2.396799742 podStartE2EDuration="5.880433964s" podCreationTimestamp="2025-12-04 12:48:27 +0000 UTC" firstStartedPulling="2025-12-04 12:48:28.826873358 +0000 UTC m=+3933.101169162" lastFinishedPulling="2025-12-04 12:48:32.31050758 +0000 UTC m=+3936.584803384" observedRunningTime="2025-12-04 12:48:32.878538853 +0000 UTC m=+3937.152834677" watchObservedRunningTime="2025-12-04 12:48:32.880433964 +0000 UTC m=+3937.154729768" Dec 04 12:48:37 crc kubenswrapper[4979]: E1204 12:48:37.547330 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice/crio-c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944\": RecentStats: unable to find data in memory cache]" Dec 04 12:48:38 crc kubenswrapper[4979]: I1204 12:48:38.097446 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:38 crc kubenswrapper[4979]: I1204 12:48:38.098014 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:38 crc kubenswrapper[4979]: I1204 12:48:38.149901 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:38 crc kubenswrapper[4979]: I1204 12:48:38.977408 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:39 crc kubenswrapper[4979]: I1204 12:48:39.024775 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nq9bk"] Dec 04 12:48:40 crc kubenswrapper[4979]: I1204 12:48:40.924999 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nq9bk" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="registry-server" containerID="cri-o://ce543bc29510b3c1595cf619094593bcb99dc7ce107829a0676ba987228a00f6" gracePeriod=2 Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.002794 4979 generic.go:334] "Generic (PLEG): container finished" podID="b752c1f8-051c-43f7-a762-080e606c76e8" containerID="ce543bc29510b3c1595cf619094593bcb99dc7ce107829a0676ba987228a00f6" exitCode=0 Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.002867 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerDied","Data":"ce543bc29510b3c1595cf619094593bcb99dc7ce107829a0676ba987228a00f6"} Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.064606 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.212543 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-utilities\") pod \"b752c1f8-051c-43f7-a762-080e606c76e8\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.212649 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-catalog-content\") pod \"b752c1f8-051c-43f7-a762-080e606c76e8\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.212681 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwbkd\" (UniqueName: \"kubernetes.io/projected/b752c1f8-051c-43f7-a762-080e606c76e8-kube-api-access-gwbkd\") pod \"b752c1f8-051c-43f7-a762-080e606c76e8\" (UID: \"b752c1f8-051c-43f7-a762-080e606c76e8\") " Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.213650 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-utilities" (OuterVolumeSpecName: "utilities") pod "b752c1f8-051c-43f7-a762-080e606c76e8" (UID: "b752c1f8-051c-43f7-a762-080e606c76e8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.219442 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b752c1f8-051c-43f7-a762-080e606c76e8-kube-api-access-gwbkd" (OuterVolumeSpecName: "kube-api-access-gwbkd") pod "b752c1f8-051c-43f7-a762-080e606c76e8" (UID: "b752c1f8-051c-43f7-a762-080e606c76e8"). InnerVolumeSpecName "kube-api-access-gwbkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.314475 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.314517 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwbkd\" (UniqueName: \"kubernetes.io/projected/b752c1f8-051c-43f7-a762-080e606c76e8-kube-api-access-gwbkd\") on node \"crc\" DevicePath \"\"" Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.331605 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b752c1f8-051c-43f7-a762-080e606c76e8" (UID: "b752c1f8-051c-43f7-a762-080e606c76e8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:48:45 crc kubenswrapper[4979]: I1204 12:48:45.416108 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b752c1f8-051c-43f7-a762-080e606c76e8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.016765 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nq9bk" event={"ID":"b752c1f8-051c-43f7-a762-080e606c76e8","Type":"ContainerDied","Data":"19b7deae8906b2a68886179f25b7c7ee6a2036ce5754f8664e142c786b671d0a"} Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.016833 4979 scope.go:117] "RemoveContainer" containerID="ce543bc29510b3c1595cf619094593bcb99dc7ce107829a0676ba987228a00f6" Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.017091 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nq9bk" Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.034329 4979 scope.go:117] "RemoveContainer" containerID="35fcaa73e19ba34fd5e867f5a7a7bfc18da9e41d5f63d38321cb0b8b0c3d3257" Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.052052 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nq9bk"] Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.058463 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nq9bk"] Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.060607 4979 scope.go:117] "RemoveContainer" containerID="3e90ecf475d711b1a15c8dcd8f9e75e939e5e51e617948222d582489f8967e8f" Dec 04 12:48:46 crc kubenswrapper[4979]: I1204 12:48:46.206897 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" path="/var/lib/kubelet/pods/b752c1f8-051c-43f7-a762-080e606c76e8/volumes" Dec 04 12:48:47 crc kubenswrapper[4979]: E1204 12:48:47.720274 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f9fae7_8a16_48da_a70a_153bdf67800f.slice/crio-c106352b3c4946ef3e35f27deae2201553222e5854110aab54c7961d4479e944\": RecentStats: unable to find data in memory cache]" Dec 04 12:48:56 crc kubenswrapper[4979]: E1204 12:48:56.236963 4979 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/208f605c3e29484a82b5b86fcb163595cc614a38089ed6e45ca9ff0638f0a2ae/diff" to get inode usage: stat /var/lib/containers/storage/overlay/208f605c3e29484a82b5b86fcb163595cc614a38089ed6e45ca9ff0638f0a2ae/diff: no such file or directory, extraDiskErr: Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.428269 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kjtft"] Dec 04 12:49:31 crc kubenswrapper[4979]: E1204 12:49:31.429338 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="extract-content" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.429353 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="extract-content" Dec 04 12:49:31 crc kubenswrapper[4979]: E1204 12:49:31.429379 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="registry-server" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.429387 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="registry-server" Dec 04 12:49:31 crc kubenswrapper[4979]: E1204 12:49:31.429410 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="extract-utilities" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.429418 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="extract-utilities" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.429602 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b752c1f8-051c-43f7-a762-080e606c76e8" containerName="registry-server" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.430922 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.446401 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjtft"] Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.563397 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-catalog-content\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.563467 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvt42\" (UniqueName: \"kubernetes.io/projected/4b90a274-ef44-4173-9dc6-01d0677bf87d-kube-api-access-pvt42\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.563502 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-utilities\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.664969 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-utilities\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.665087 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-catalog-content\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.665144 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvt42\" (UniqueName: \"kubernetes.io/projected/4b90a274-ef44-4173-9dc6-01d0677bf87d-kube-api-access-pvt42\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.665550 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-utilities\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.665571 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-catalog-content\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.693263 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvt42\" (UniqueName: \"kubernetes.io/projected/4b90a274-ef44-4173-9dc6-01d0677bf87d-kube-api-access-pvt42\") pod \"redhat-marketplace-kjtft\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:31 crc kubenswrapper[4979]: I1204 12:49:31.750374 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:32 crc kubenswrapper[4979]: I1204 12:49:32.208848 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjtft"] Dec 04 12:49:32 crc kubenswrapper[4979]: I1204 12:49:32.361554 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjtft" event={"ID":"4b90a274-ef44-4173-9dc6-01d0677bf87d","Type":"ContainerStarted","Data":"8dc28e71032dd6426d60ade702dae6b32c40aa9ab3993e0c3fab4c33ee9db3cc"} Dec 04 12:49:33 crc kubenswrapper[4979]: I1204 12:49:33.370343 4979 generic.go:334] "Generic (PLEG): container finished" podID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerID="4e9c465e7c366c3eee7b66caea88ce660d84c851748e3f0b3ecc4a38ec7fb19f" exitCode=0 Dec 04 12:49:33 crc kubenswrapper[4979]: I1204 12:49:33.370384 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjtft" event={"ID":"4b90a274-ef44-4173-9dc6-01d0677bf87d","Type":"ContainerDied","Data":"4e9c465e7c366c3eee7b66caea88ce660d84c851748e3f0b3ecc4a38ec7fb19f"} Dec 04 12:49:33 crc kubenswrapper[4979]: I1204 12:49:33.373429 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:49:35 crc kubenswrapper[4979]: I1204 12:49:35.387808 4979 generic.go:334] "Generic (PLEG): container finished" podID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerID="9cb887b011a68eb05c13001b49556d8172531a68a3e541896fe6f2010d78f844" exitCode=0 Dec 04 12:49:35 crc kubenswrapper[4979]: I1204 12:49:35.387874 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjtft" event={"ID":"4b90a274-ef44-4173-9dc6-01d0677bf87d","Type":"ContainerDied","Data":"9cb887b011a68eb05c13001b49556d8172531a68a3e541896fe6f2010d78f844"} Dec 04 12:49:37 crc kubenswrapper[4979]: I1204 12:49:37.405955 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjtft" event={"ID":"4b90a274-ef44-4173-9dc6-01d0677bf87d","Type":"ContainerStarted","Data":"c07dc6d76b7832d533912c67debb60a3acd080295e4702a6f283ea6f4b49bbbd"} Dec 04 12:49:37 crc kubenswrapper[4979]: I1204 12:49:37.428670 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kjtft" podStartSLOduration=3.95887226 podStartE2EDuration="6.42864921s" podCreationTimestamp="2025-12-04 12:49:31 +0000 UTC" firstStartedPulling="2025-12-04 12:49:33.373096609 +0000 UTC m=+3997.647392413" lastFinishedPulling="2025-12-04 12:49:35.842873559 +0000 UTC m=+4000.117169363" observedRunningTime="2025-12-04 12:49:37.422363719 +0000 UTC m=+4001.696659543" watchObservedRunningTime="2025-12-04 12:49:37.42864921 +0000 UTC m=+4001.702945014" Dec 04 12:49:41 crc kubenswrapper[4979]: I1204 12:49:41.751085 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:41 crc kubenswrapper[4979]: I1204 12:49:41.751580 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:41 crc kubenswrapper[4979]: I1204 12:49:41.806883 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:42 crc kubenswrapper[4979]: I1204 12:49:42.483941 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:42 crc kubenswrapper[4979]: I1204 12:49:42.531988 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjtft"] Dec 04 12:49:44 crc kubenswrapper[4979]: I1204 12:49:44.459633 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kjtft" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="registry-server" containerID="cri-o://c07dc6d76b7832d533912c67debb60a3acd080295e4702a6f283ea6f4b49bbbd" gracePeriod=2 Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.474926 4979 generic.go:334] "Generic (PLEG): container finished" podID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerID="c07dc6d76b7832d533912c67debb60a3acd080295e4702a6f283ea6f4b49bbbd" exitCode=0 Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.475173 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjtft" event={"ID":"4b90a274-ef44-4173-9dc6-01d0677bf87d","Type":"ContainerDied","Data":"c07dc6d76b7832d533912c67debb60a3acd080295e4702a6f283ea6f4b49bbbd"} Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.722101 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.780941 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvt42\" (UniqueName: \"kubernetes.io/projected/4b90a274-ef44-4173-9dc6-01d0677bf87d-kube-api-access-pvt42\") pod \"4b90a274-ef44-4173-9dc6-01d0677bf87d\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.781020 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-utilities\") pod \"4b90a274-ef44-4173-9dc6-01d0677bf87d\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.781050 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-catalog-content\") pod \"4b90a274-ef44-4173-9dc6-01d0677bf87d\" (UID: \"4b90a274-ef44-4173-9dc6-01d0677bf87d\") " Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.782626 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-utilities" (OuterVolumeSpecName: "utilities") pod "4b90a274-ef44-4173-9dc6-01d0677bf87d" (UID: "4b90a274-ef44-4173-9dc6-01d0677bf87d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.786656 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b90a274-ef44-4173-9dc6-01d0677bf87d-kube-api-access-pvt42" (OuterVolumeSpecName: "kube-api-access-pvt42") pod "4b90a274-ef44-4173-9dc6-01d0677bf87d" (UID: "4b90a274-ef44-4173-9dc6-01d0677bf87d"). InnerVolumeSpecName "kube-api-access-pvt42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.801277 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4b90a274-ef44-4173-9dc6-01d0677bf87d" (UID: "4b90a274-ef44-4173-9dc6-01d0677bf87d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.882853 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.882886 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4b90a274-ef44-4173-9dc6-01d0677bf87d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:49:45 crc kubenswrapper[4979]: I1204 12:49:45.882897 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvt42\" (UniqueName: \"kubernetes.io/projected/4b90a274-ef44-4173-9dc6-01d0677bf87d-kube-api-access-pvt42\") on node \"crc\" DevicePath \"\"" Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.484019 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjtft" event={"ID":"4b90a274-ef44-4173-9dc6-01d0677bf87d","Type":"ContainerDied","Data":"8dc28e71032dd6426d60ade702dae6b32c40aa9ab3993e0c3fab4c33ee9db3cc"} Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.484097 4979 scope.go:117] "RemoveContainer" containerID="c07dc6d76b7832d533912c67debb60a3acd080295e4702a6f283ea6f4b49bbbd" Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.484389 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjtft" Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.511366 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjtft"] Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.514070 4979 scope.go:117] "RemoveContainer" containerID="9cb887b011a68eb05c13001b49556d8172531a68a3e541896fe6f2010d78f844" Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.517950 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjtft"] Dec 04 12:49:46 crc kubenswrapper[4979]: I1204 12:49:46.536199 4979 scope.go:117] "RemoveContainer" containerID="4e9c465e7c366c3eee7b66caea88ce660d84c851748e3f0b3ecc4a38ec7fb19f" Dec 04 12:49:48 crc kubenswrapper[4979]: I1204 12:49:48.207826 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" path="/var/lib/kubelet/pods/4b90a274-ef44-4173-9dc6-01d0677bf87d/volumes" Dec 04 12:50:28 crc kubenswrapper[4979]: I1204 12:50:28.040782 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:50:28 crc kubenswrapper[4979]: I1204 12:50:28.041353 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:50:58 crc kubenswrapper[4979]: I1204 12:50:58.040323 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:50:58 crc kubenswrapper[4979]: I1204 12:50:58.040871 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.040406 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.041677 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.041749 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.042727 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c227c4073a3b2038c807150c1376009b949d75a319f0bf4a7c3711918aa7c2d"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.042803 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://3c227c4073a3b2038c807150c1376009b949d75a319f0bf4a7c3711918aa7c2d" gracePeriod=600 Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.275109 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="3c227c4073a3b2038c807150c1376009b949d75a319f0bf4a7c3711918aa7c2d" exitCode=0 Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.275191 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"3c227c4073a3b2038c807150c1376009b949d75a319f0bf4a7c3711918aa7c2d"} Dec 04 12:51:28 crc kubenswrapper[4979]: I1204 12:51:28.275586 4979 scope.go:117] "RemoveContainer" containerID="aab129513c86db7e490c5d4d5ce5d79d4b989d36f427f0ee6aa83c3875ecd2e0" Dec 04 12:51:29 crc kubenswrapper[4979]: I1204 12:51:29.285417 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7"} Dec 04 12:53:28 crc kubenswrapper[4979]: I1204 12:53:28.040815 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:53:28 crc kubenswrapper[4979]: I1204 12:53:28.041515 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:53:58 crc kubenswrapper[4979]: I1204 12:53:58.041223 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:53:58 crc kubenswrapper[4979]: I1204 12:53:58.041831 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.040936 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.042154 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.042219 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.042758 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.042814 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" gracePeriod=600 Dec 04 12:54:28 crc kubenswrapper[4979]: E1204 12:54:28.173843 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.715524 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" exitCode=0 Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.715579 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7"} Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.715827 4979 scope.go:117] "RemoveContainer" containerID="3c227c4073a3b2038c807150c1376009b949d75a319f0bf4a7c3711918aa7c2d" Dec 04 12:54:28 crc kubenswrapper[4979]: I1204 12:54:28.716216 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:54:28 crc kubenswrapper[4979]: E1204 12:54:28.716438 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:54:41 crc kubenswrapper[4979]: I1204 12:54:41.199409 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:54:41 crc kubenswrapper[4979]: E1204 12:54:41.200167 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:54:56 crc kubenswrapper[4979]: I1204 12:54:56.207505 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:54:56 crc kubenswrapper[4979]: E1204 12:54:56.208145 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.245060 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gx6vm"] Dec 04 12:54:57 crc kubenswrapper[4979]: E1204 12:54:57.245498 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="extract-content" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.245516 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="extract-content" Dec 04 12:54:57 crc kubenswrapper[4979]: E1204 12:54:57.245536 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="extract-utilities" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.245544 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="extract-utilities" Dec 04 12:54:57 crc kubenswrapper[4979]: E1204 12:54:57.245578 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="registry-server" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.245586 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="registry-server" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.245761 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b90a274-ef44-4173-9dc6-01d0677bf87d" containerName="registry-server" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.247118 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.259506 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gx6vm"] Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.371178 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26knn\" (UniqueName: \"kubernetes.io/projected/a9f4c92e-af0f-4121-9a65-059928ef57b8-kube-api-access-26knn\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.371268 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-catalog-content\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.371335 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-utilities\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.472317 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26knn\" (UniqueName: \"kubernetes.io/projected/a9f4c92e-af0f-4121-9a65-059928ef57b8-kube-api-access-26knn\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.472391 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-catalog-content\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.472432 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-utilities\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.472964 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-utilities\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.473028 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-catalog-content\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.872244 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26knn\" (UniqueName: \"kubernetes.io/projected/a9f4c92e-af0f-4121-9a65-059928ef57b8-kube-api-access-26knn\") pod \"certified-operators-gx6vm\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:57 crc kubenswrapper[4979]: I1204 12:54:57.874206 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:54:58 crc kubenswrapper[4979]: I1204 12:54:58.321925 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gx6vm"] Dec 04 12:54:58 crc kubenswrapper[4979]: I1204 12:54:58.942406 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx6vm" event={"ID":"a9f4c92e-af0f-4121-9a65-059928ef57b8","Type":"ContainerStarted","Data":"a58e5a877682ba72e5e37a2f360cb109c2dcca37fd541149a45843dc837e346e"} Dec 04 12:54:59 crc kubenswrapper[4979]: I1204 12:54:59.956487 4979 generic.go:334] "Generic (PLEG): container finished" podID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerID="898ae3ec4f09957db291e8be56f4a34574094a74e862c2a37e4066679e5a76fd" exitCode=0 Dec 04 12:54:59 crc kubenswrapper[4979]: I1204 12:54:59.956592 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx6vm" event={"ID":"a9f4c92e-af0f-4121-9a65-059928ef57b8","Type":"ContainerDied","Data":"898ae3ec4f09957db291e8be56f4a34574094a74e862c2a37e4066679e5a76fd"} Dec 04 12:54:59 crc kubenswrapper[4979]: I1204 12:54:59.959162 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 12:55:02 crc kubenswrapper[4979]: I1204 12:55:02.979588 4979 generic.go:334] "Generic (PLEG): container finished" podID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerID="0744c94576185ebc7552d36dc109dd0e2f3ec83c0bae9a7febec5995488d5482" exitCode=0 Dec 04 12:55:02 crc kubenswrapper[4979]: I1204 12:55:02.979703 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx6vm" event={"ID":"a9f4c92e-af0f-4121-9a65-059928ef57b8","Type":"ContainerDied","Data":"0744c94576185ebc7552d36dc109dd0e2f3ec83c0bae9a7febec5995488d5482"} Dec 04 12:55:07 crc kubenswrapper[4979]: I1204 12:55:07.009635 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx6vm" event={"ID":"a9f4c92e-af0f-4121-9a65-059928ef57b8","Type":"ContainerStarted","Data":"b2594e35a6a5acf3bd59d4e4e36bac0914e85ba17700765461005aad956b64b1"} Dec 04 12:55:08 crc kubenswrapper[4979]: I1204 12:55:08.038120 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gx6vm" podStartSLOduration=4.55590081 podStartE2EDuration="11.038103168s" podCreationTimestamp="2025-12-04 12:54:57 +0000 UTC" firstStartedPulling="2025-12-04 12:54:59.95891511 +0000 UTC m=+4324.233210914" lastFinishedPulling="2025-12-04 12:55:06.441117428 +0000 UTC m=+4330.715413272" observedRunningTime="2025-12-04 12:55:08.033694868 +0000 UTC m=+4332.307990702" watchObservedRunningTime="2025-12-04 12:55:08.038103168 +0000 UTC m=+4332.312398972" Dec 04 12:55:10 crc kubenswrapper[4979]: I1204 12:55:10.199097 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:55:10 crc kubenswrapper[4979]: E1204 12:55:10.199994 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:55:17 crc kubenswrapper[4979]: I1204 12:55:17.874818 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:55:17 crc kubenswrapper[4979]: I1204 12:55:17.875830 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:55:17 crc kubenswrapper[4979]: I1204 12:55:17.924938 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:55:18 crc kubenswrapper[4979]: I1204 12:55:18.143715 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:55:18 crc kubenswrapper[4979]: I1204 12:55:18.191848 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gx6vm"] Dec 04 12:55:20 crc kubenswrapper[4979]: I1204 12:55:20.110518 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gx6vm" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="registry-server" containerID="cri-o://b2594e35a6a5acf3bd59d4e4e36bac0914e85ba17700765461005aad956b64b1" gracePeriod=2 Dec 04 12:55:21 crc kubenswrapper[4979]: I1204 12:55:21.125951 4979 generic.go:334] "Generic (PLEG): container finished" podID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerID="b2594e35a6a5acf3bd59d4e4e36bac0914e85ba17700765461005aad956b64b1" exitCode=0 Dec 04 12:55:21 crc kubenswrapper[4979]: I1204 12:55:21.126027 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx6vm" event={"ID":"a9f4c92e-af0f-4121-9a65-059928ef57b8","Type":"ContainerDied","Data":"b2594e35a6a5acf3bd59d4e4e36bac0914e85ba17700765461005aad956b64b1"} Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.200137 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:55:22 crc kubenswrapper[4979]: E1204 12:55:22.200791 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.281002 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.378448 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-utilities\") pod \"a9f4c92e-af0f-4121-9a65-059928ef57b8\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.378544 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26knn\" (UniqueName: \"kubernetes.io/projected/a9f4c92e-af0f-4121-9a65-059928ef57b8-kube-api-access-26knn\") pod \"a9f4c92e-af0f-4121-9a65-059928ef57b8\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.378643 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-catalog-content\") pod \"a9f4c92e-af0f-4121-9a65-059928ef57b8\" (UID: \"a9f4c92e-af0f-4121-9a65-059928ef57b8\") " Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.379573 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-utilities" (OuterVolumeSpecName: "utilities") pod "a9f4c92e-af0f-4121-9a65-059928ef57b8" (UID: "a9f4c92e-af0f-4121-9a65-059928ef57b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.383978 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9f4c92e-af0f-4121-9a65-059928ef57b8-kube-api-access-26knn" (OuterVolumeSpecName: "kube-api-access-26knn") pod "a9f4c92e-af0f-4121-9a65-059928ef57b8" (UID: "a9f4c92e-af0f-4121-9a65-059928ef57b8"). InnerVolumeSpecName "kube-api-access-26knn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.429913 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a9f4c92e-af0f-4121-9a65-059928ef57b8" (UID: "a9f4c92e-af0f-4121-9a65-059928ef57b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.479981 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.480188 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26knn\" (UniqueName: \"kubernetes.io/projected/a9f4c92e-af0f-4121-9a65-059928ef57b8-kube-api-access-26knn\") on node \"crc\" DevicePath \"\"" Dec 04 12:55:22 crc kubenswrapper[4979]: I1204 12:55:22.480277 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a9f4c92e-af0f-4121-9a65-059928ef57b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.143781 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gx6vm" event={"ID":"a9f4c92e-af0f-4121-9a65-059928ef57b8","Type":"ContainerDied","Data":"a58e5a877682ba72e5e37a2f360cb109c2dcca37fd541149a45843dc837e346e"} Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.143858 4979 scope.go:117] "RemoveContainer" containerID="b2594e35a6a5acf3bd59d4e4e36bac0914e85ba17700765461005aad956b64b1" Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.144107 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gx6vm" Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.185444 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gx6vm"] Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.188136 4979 scope.go:117] "RemoveContainer" containerID="0744c94576185ebc7552d36dc109dd0e2f3ec83c0bae9a7febec5995488d5482" Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.193494 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gx6vm"] Dec 04 12:55:23 crc kubenswrapper[4979]: I1204 12:55:23.209476 4979 scope.go:117] "RemoveContainer" containerID="898ae3ec4f09957db291e8be56f4a34574094a74e862c2a37e4066679e5a76fd" Dec 04 12:55:24 crc kubenswrapper[4979]: I1204 12:55:24.209352 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" path="/var/lib/kubelet/pods/a9f4c92e-af0f-4121-9a65-059928ef57b8/volumes" Dec 04 12:55:36 crc kubenswrapper[4979]: I1204 12:55:36.204862 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:55:36 crc kubenswrapper[4979]: E1204 12:55:36.206724 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:55:48 crc kubenswrapper[4979]: I1204 12:55:48.199733 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:55:48 crc kubenswrapper[4979]: E1204 12:55:48.200466 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:56:03 crc kubenswrapper[4979]: I1204 12:56:03.198683 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:56:03 crc kubenswrapper[4979]: E1204 12:56:03.199272 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:56:18 crc kubenswrapper[4979]: I1204 12:56:18.199534 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:56:18 crc kubenswrapper[4979]: E1204 12:56:18.200590 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:56:29 crc kubenswrapper[4979]: I1204 12:56:29.198933 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:56:29 crc kubenswrapper[4979]: E1204 12:56:29.200436 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:56:44 crc kubenswrapper[4979]: I1204 12:56:44.199056 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:56:44 crc kubenswrapper[4979]: E1204 12:56:44.199867 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:56:58 crc kubenswrapper[4979]: I1204 12:56:58.198765 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:56:58 crc kubenswrapper[4979]: E1204 12:56:58.199679 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:57:10 crc kubenswrapper[4979]: I1204 12:57:10.199478 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:57:10 crc kubenswrapper[4979]: E1204 12:57:10.200647 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:57:24 crc kubenswrapper[4979]: I1204 12:57:24.199572 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:57:24 crc kubenswrapper[4979]: E1204 12:57:24.200510 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:57:36 crc kubenswrapper[4979]: I1204 12:57:36.203697 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:57:36 crc kubenswrapper[4979]: E1204 12:57:36.204627 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.198484 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:57:48 crc kubenswrapper[4979]: E1204 12:57:48.199211 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.802011 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-f2jlf"] Dec 04 12:57:48 crc kubenswrapper[4979]: E1204 12:57:48.802684 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="extract-content" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.802704 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="extract-content" Dec 04 12:57:48 crc kubenswrapper[4979]: E1204 12:57:48.802733 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="registry-server" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.802741 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="registry-server" Dec 04 12:57:48 crc kubenswrapper[4979]: E1204 12:57:48.802750 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="extract-utilities" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.802756 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="extract-utilities" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.802900 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9f4c92e-af0f-4121-9a65-059928ef57b8" containerName="registry-server" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.803932 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.813058 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2jlf"] Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.875046 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-catalog-content\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.875122 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-utilities\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.875148 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9xnx\" (UniqueName: \"kubernetes.io/projected/646cfb90-ed81-484f-a71f-df48e47c6238-kube-api-access-w9xnx\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.976830 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-catalog-content\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.976900 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-utilities\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.976918 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9xnx\" (UniqueName: \"kubernetes.io/projected/646cfb90-ed81-484f-a71f-df48e47c6238-kube-api-access-w9xnx\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.977536 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-catalog-content\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.977539 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-utilities\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:48 crc kubenswrapper[4979]: I1204 12:57:48.995957 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9xnx\" (UniqueName: \"kubernetes.io/projected/646cfb90-ed81-484f-a71f-df48e47c6238-kube-api-access-w9xnx\") pod \"community-operators-f2jlf\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:49 crc kubenswrapper[4979]: I1204 12:57:49.123942 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:49 crc kubenswrapper[4979]: I1204 12:57:49.412877 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-f2jlf"] Dec 04 12:57:50 crc kubenswrapper[4979]: I1204 12:57:50.232527 4979 generic.go:334] "Generic (PLEG): container finished" podID="646cfb90-ed81-484f-a71f-df48e47c6238" containerID="349699b71f82b6c88289ae3fe243a9f116b4ad88340c9e8e07d11179ec4b8137" exitCode=0 Dec 04 12:57:50 crc kubenswrapper[4979]: I1204 12:57:50.232595 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2jlf" event={"ID":"646cfb90-ed81-484f-a71f-df48e47c6238","Type":"ContainerDied","Data":"349699b71f82b6c88289ae3fe243a9f116b4ad88340c9e8e07d11179ec4b8137"} Dec 04 12:57:50 crc kubenswrapper[4979]: I1204 12:57:50.232853 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2jlf" event={"ID":"646cfb90-ed81-484f-a71f-df48e47c6238","Type":"ContainerStarted","Data":"8fb9746b508a579fae05db7c521dd47491e74a08c3210afa32dfc186930716cc"} Dec 04 12:57:52 crc kubenswrapper[4979]: I1204 12:57:52.251165 4979 generic.go:334] "Generic (PLEG): container finished" podID="646cfb90-ed81-484f-a71f-df48e47c6238" containerID="c764cae6a80891848d69471feced2fac80747024da68fdb00ea68825038034fd" exitCode=0 Dec 04 12:57:52 crc kubenswrapper[4979]: I1204 12:57:52.251201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2jlf" event={"ID":"646cfb90-ed81-484f-a71f-df48e47c6238","Type":"ContainerDied","Data":"c764cae6a80891848d69471feced2fac80747024da68fdb00ea68825038034fd"} Dec 04 12:57:53 crc kubenswrapper[4979]: I1204 12:57:53.259877 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2jlf" event={"ID":"646cfb90-ed81-484f-a71f-df48e47c6238","Type":"ContainerStarted","Data":"d377b7c4446c219c72e6f87a456382e8032009f943cedac5c01c0eeefd234bfb"} Dec 04 12:57:53 crc kubenswrapper[4979]: I1204 12:57:53.283601 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-f2jlf" podStartSLOduration=2.842195959 podStartE2EDuration="5.283582853s" podCreationTimestamp="2025-12-04 12:57:48 +0000 UTC" firstStartedPulling="2025-12-04 12:57:50.235243757 +0000 UTC m=+4494.509539561" lastFinishedPulling="2025-12-04 12:57:52.676630651 +0000 UTC m=+4496.950926455" observedRunningTime="2025-12-04 12:57:53.276370208 +0000 UTC m=+4497.550666012" watchObservedRunningTime="2025-12-04 12:57:53.283582853 +0000 UTC m=+4497.557878657" Dec 04 12:57:59 crc kubenswrapper[4979]: I1204 12:57:59.124475 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:59 crc kubenswrapper[4979]: I1204 12:57:59.124853 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:59 crc kubenswrapper[4979]: I1204 12:57:59.165237 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:59 crc kubenswrapper[4979]: I1204 12:57:59.336804 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:57:59 crc kubenswrapper[4979]: I1204 12:57:59.396329 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2jlf"] Dec 04 12:58:00 crc kubenswrapper[4979]: I1204 12:58:00.198660 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:58:00 crc kubenswrapper[4979]: E1204 12:58:00.198928 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:58:01 crc kubenswrapper[4979]: I1204 12:58:01.310708 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-f2jlf" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="registry-server" containerID="cri-o://d377b7c4446c219c72e6f87a456382e8032009f943cedac5c01c0eeefd234bfb" gracePeriod=2 Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.322147 4979 generic.go:334] "Generic (PLEG): container finished" podID="646cfb90-ed81-484f-a71f-df48e47c6238" containerID="d377b7c4446c219c72e6f87a456382e8032009f943cedac5c01c0eeefd234bfb" exitCode=0 Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.322201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2jlf" event={"ID":"646cfb90-ed81-484f-a71f-df48e47c6238","Type":"ContainerDied","Data":"d377b7c4446c219c72e6f87a456382e8032009f943cedac5c01c0eeefd234bfb"} Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.540635 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.689424 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-utilities\") pod \"646cfb90-ed81-484f-a71f-df48e47c6238\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.689780 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-catalog-content\") pod \"646cfb90-ed81-484f-a71f-df48e47c6238\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.689913 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9xnx\" (UniqueName: \"kubernetes.io/projected/646cfb90-ed81-484f-a71f-df48e47c6238-kube-api-access-w9xnx\") pod \"646cfb90-ed81-484f-a71f-df48e47c6238\" (UID: \"646cfb90-ed81-484f-a71f-df48e47c6238\") " Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.691099 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-utilities" (OuterVolumeSpecName: "utilities") pod "646cfb90-ed81-484f-a71f-df48e47c6238" (UID: "646cfb90-ed81-484f-a71f-df48e47c6238"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.695585 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/646cfb90-ed81-484f-a71f-df48e47c6238-kube-api-access-w9xnx" (OuterVolumeSpecName: "kube-api-access-w9xnx") pod "646cfb90-ed81-484f-a71f-df48e47c6238" (UID: "646cfb90-ed81-484f-a71f-df48e47c6238"). InnerVolumeSpecName "kube-api-access-w9xnx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.737328 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "646cfb90-ed81-484f-a71f-df48e47c6238" (UID: "646cfb90-ed81-484f-a71f-df48e47c6238"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.791759 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.792111 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9xnx\" (UniqueName: \"kubernetes.io/projected/646cfb90-ed81-484f-a71f-df48e47c6238-kube-api-access-w9xnx\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:02 crc kubenswrapper[4979]: I1204 12:58:02.792355 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/646cfb90-ed81-484f-a71f-df48e47c6238-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.331814 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-f2jlf" event={"ID":"646cfb90-ed81-484f-a71f-df48e47c6238","Type":"ContainerDied","Data":"8fb9746b508a579fae05db7c521dd47491e74a08c3210afa32dfc186930716cc"} Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.331867 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-f2jlf" Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.332435 4979 scope.go:117] "RemoveContainer" containerID="d377b7c4446c219c72e6f87a456382e8032009f943cedac5c01c0eeefd234bfb" Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.349922 4979 scope.go:117] "RemoveContainer" containerID="c764cae6a80891848d69471feced2fac80747024da68fdb00ea68825038034fd" Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.365858 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-f2jlf"] Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.371203 4979 scope.go:117] "RemoveContainer" containerID="349699b71f82b6c88289ae3fe243a9f116b4ad88340c9e8e07d11179ec4b8137" Dec 04 12:58:03 crc kubenswrapper[4979]: I1204 12:58:03.371855 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-f2jlf"] Dec 04 12:58:04 crc kubenswrapper[4979]: I1204 12:58:04.213195 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" path="/var/lib/kubelet/pods/646cfb90-ed81-484f-a71f-df48e47c6238/volumes" Dec 04 12:58:15 crc kubenswrapper[4979]: I1204 12:58:15.199125 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:58:15 crc kubenswrapper[4979]: E1204 12:58:15.199839 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:58:26 crc kubenswrapper[4979]: I1204 12:58:26.203446 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:58:26 crc kubenswrapper[4979]: E1204 12:58:26.204161 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.605814 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xrpdj"] Dec 04 12:58:31 crc kubenswrapper[4979]: E1204 12:58:31.606667 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="extract-content" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.606682 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="extract-content" Dec 04 12:58:31 crc kubenswrapper[4979]: E1204 12:58:31.606696 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="registry-server" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.606703 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="registry-server" Dec 04 12:58:31 crc kubenswrapper[4979]: E1204 12:58:31.606724 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="extract-utilities" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.606731 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="extract-utilities" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.606870 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="646cfb90-ed81-484f-a71f-df48e47c6238" containerName="registry-server" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.607989 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.617827 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xrpdj"] Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.695400 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-utilities\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.695446 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5g5c\" (UniqueName: \"kubernetes.io/projected/a4ea5ec8-b23f-4046-aa26-57f073f4e078-kube-api-access-z5g5c\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.695472 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-catalog-content\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.796650 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-utilities\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.796967 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5g5c\" (UniqueName: \"kubernetes.io/projected/a4ea5ec8-b23f-4046-aa26-57f073f4e078-kube-api-access-z5g5c\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.797103 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-catalog-content\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.797238 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-utilities\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.797641 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-catalog-content\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.816554 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5g5c\" (UniqueName: \"kubernetes.io/projected/a4ea5ec8-b23f-4046-aa26-57f073f4e078-kube-api-access-z5g5c\") pod \"redhat-operators-xrpdj\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:31 crc kubenswrapper[4979]: I1204 12:58:31.930952 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:32 crc kubenswrapper[4979]: I1204 12:58:32.375643 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xrpdj"] Dec 04 12:58:32 crc kubenswrapper[4979]: I1204 12:58:32.530135 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerStarted","Data":"bac16aa7c64d8f41244edf5811db146d2dc254d6e4c2b52aafd133974d11de12"} Dec 04 12:58:33 crc kubenswrapper[4979]: I1204 12:58:33.542268 4979 generic.go:334] "Generic (PLEG): container finished" podID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerID="0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b" exitCode=0 Dec 04 12:58:33 crc kubenswrapper[4979]: I1204 12:58:33.542347 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerDied","Data":"0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b"} Dec 04 12:58:34 crc kubenswrapper[4979]: I1204 12:58:34.551583 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerStarted","Data":"157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174"} Dec 04 12:58:35 crc kubenswrapper[4979]: I1204 12:58:35.561912 4979 generic.go:334] "Generic (PLEG): container finished" podID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerID="157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174" exitCode=0 Dec 04 12:58:35 crc kubenswrapper[4979]: I1204 12:58:35.562006 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerDied","Data":"157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174"} Dec 04 12:58:36 crc kubenswrapper[4979]: I1204 12:58:36.576074 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerStarted","Data":"f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13"} Dec 04 12:58:36 crc kubenswrapper[4979]: I1204 12:58:36.603216 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xrpdj" podStartSLOduration=3.164699658 podStartE2EDuration="5.603180275s" podCreationTimestamp="2025-12-04 12:58:31 +0000 UTC" firstStartedPulling="2025-12-04 12:58:33.545925266 +0000 UTC m=+4537.820221070" lastFinishedPulling="2025-12-04 12:58:35.984405883 +0000 UTC m=+4540.258701687" observedRunningTime="2025-12-04 12:58:36.592345191 +0000 UTC m=+4540.866641015" watchObservedRunningTime="2025-12-04 12:58:36.603180275 +0000 UTC m=+4540.877476079" Dec 04 12:58:41 crc kubenswrapper[4979]: I1204 12:58:41.198949 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:58:41 crc kubenswrapper[4979]: E1204 12:58:41.199547 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:58:41 crc kubenswrapper[4979]: I1204 12:58:41.931095 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:41 crc kubenswrapper[4979]: I1204 12:58:41.931153 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:41 crc kubenswrapper[4979]: I1204 12:58:41.971753 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:42 crc kubenswrapper[4979]: I1204 12:58:42.661747 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:42 crc kubenswrapper[4979]: I1204 12:58:42.714316 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xrpdj"] Dec 04 12:58:44 crc kubenswrapper[4979]: I1204 12:58:44.628123 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xrpdj" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="registry-server" containerID="cri-o://f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13" gracePeriod=2 Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.031184 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.080392 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-utilities\") pod \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.080520 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5g5c\" (UniqueName: \"kubernetes.io/projected/a4ea5ec8-b23f-4046-aa26-57f073f4e078-kube-api-access-z5g5c\") pod \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.080569 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-catalog-content\") pod \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\" (UID: \"a4ea5ec8-b23f-4046-aa26-57f073f4e078\") " Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.081933 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-utilities" (OuterVolumeSpecName: "utilities") pod "a4ea5ec8-b23f-4046-aa26-57f073f4e078" (UID: "a4ea5ec8-b23f-4046-aa26-57f073f4e078"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.085687 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ea5ec8-b23f-4046-aa26-57f073f4e078-kube-api-access-z5g5c" (OuterVolumeSpecName: "kube-api-access-z5g5c") pod "a4ea5ec8-b23f-4046-aa26-57f073f4e078" (UID: "a4ea5ec8-b23f-4046-aa26-57f073f4e078"). InnerVolumeSpecName "kube-api-access-z5g5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.182582 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5g5c\" (UniqueName: \"kubernetes.io/projected/a4ea5ec8-b23f-4046-aa26-57f073f4e078-kube-api-access-z5g5c\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.182611 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.638103 4979 generic.go:334] "Generic (PLEG): container finished" podID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerID="f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13" exitCode=0 Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.638162 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xrpdj" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.638154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerDied","Data":"f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13"} Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.638845 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xrpdj" event={"ID":"a4ea5ec8-b23f-4046-aa26-57f073f4e078","Type":"ContainerDied","Data":"bac16aa7c64d8f41244edf5811db146d2dc254d6e4c2b52aafd133974d11de12"} Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.638870 4979 scope.go:117] "RemoveContainer" containerID="f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.657882 4979 scope.go:117] "RemoveContainer" containerID="157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.686491 4979 scope.go:117] "RemoveContainer" containerID="0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.711218 4979 scope.go:117] "RemoveContainer" containerID="f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13" Dec 04 12:58:45 crc kubenswrapper[4979]: E1204 12:58:45.711944 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13\": container with ID starting with f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13 not found: ID does not exist" containerID="f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.711979 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13"} err="failed to get container status \"f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13\": rpc error: code = NotFound desc = could not find container \"f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13\": container with ID starting with f93adc6edbc8130f366658230ac7e6512e4b19385223b4397cada2199d602a13 not found: ID does not exist" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.712006 4979 scope.go:117] "RemoveContainer" containerID="157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174" Dec 04 12:58:45 crc kubenswrapper[4979]: E1204 12:58:45.712324 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174\": container with ID starting with 157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174 not found: ID does not exist" containerID="157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.712345 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174"} err="failed to get container status \"157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174\": rpc error: code = NotFound desc = could not find container \"157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174\": container with ID starting with 157c894fa614208eb7c21d34624e43c96de1d6932e7acbe20a7c3d313d486174 not found: ID does not exist" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.712360 4979 scope.go:117] "RemoveContainer" containerID="0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b" Dec 04 12:58:45 crc kubenswrapper[4979]: E1204 12:58:45.712577 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b\": container with ID starting with 0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b not found: ID does not exist" containerID="0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b" Dec 04 12:58:45 crc kubenswrapper[4979]: I1204 12:58:45.712595 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b"} err="failed to get container status \"0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b\": rpc error: code = NotFound desc = could not find container \"0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b\": container with ID starting with 0cc0d7b7275f7d60237e536e86cde7fa1004e1f0eb4e534e68e9f033cc63ad2b not found: ID does not exist" Dec 04 12:58:46 crc kubenswrapper[4979]: I1204 12:58:46.161094 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4ea5ec8-b23f-4046-aa26-57f073f4e078" (UID: "a4ea5ec8-b23f-4046-aa26-57f073f4e078"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 12:58:46 crc kubenswrapper[4979]: I1204 12:58:46.194678 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4ea5ec8-b23f-4046-aa26-57f073f4e078-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:46 crc kubenswrapper[4979]: I1204 12:58:46.263117 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xrpdj"] Dec 04 12:58:46 crc kubenswrapper[4979]: I1204 12:58:46.271602 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xrpdj"] Dec 04 12:58:48 crc kubenswrapper[4979]: I1204 12:58:48.207134 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" path="/var/lib/kubelet/pods/a4ea5ec8-b23f-4046-aa26-57f073f4e078/volumes" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.458349 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-j692p"] Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.464688 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-j692p"] Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.629920 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-f77rl"] Dec 04 12:58:53 crc kubenswrapper[4979]: E1204 12:58:53.630193 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="registry-server" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.630212 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="registry-server" Dec 04 12:58:53 crc kubenswrapper[4979]: E1204 12:58:53.630245 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="extract-content" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.630252 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="extract-content" Dec 04 12:58:53 crc kubenswrapper[4979]: E1204 12:58:53.630264 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="extract-utilities" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.630272 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="extract-utilities" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.630429 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ea5ec8-b23f-4046-aa26-57f073f4e078" containerName="registry-server" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.630880 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.633775 4979 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-tm2n7" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.633791 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.633996 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.634141 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.643994 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-f77rl"] Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.692966 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22kgn\" (UniqueName: \"kubernetes.io/projected/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-kube-api-access-22kgn\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.693314 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-crc-storage\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.693572 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-node-mnt\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.795255 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-node-mnt\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.795332 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22kgn\" (UniqueName: \"kubernetes.io/projected/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-kube-api-access-22kgn\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.795361 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-crc-storage\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.795939 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-node-mnt\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.796048 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-crc-storage\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.815255 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22kgn\" (UniqueName: \"kubernetes.io/projected/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-kube-api-access-22kgn\") pod \"crc-storage-crc-f77rl\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:53 crc kubenswrapper[4979]: I1204 12:58:53.951741 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:54 crc kubenswrapper[4979]: I1204 12:58:54.207260 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dc1c6f5-64eb-456b-b7ff-6f951b425df4" path="/var/lib/kubelet/pods/8dc1c6f5-64eb-456b-b7ff-6f951b425df4/volumes" Dec 04 12:58:54 crc kubenswrapper[4979]: I1204 12:58:54.382079 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-f77rl"] Dec 04 12:58:54 crc kubenswrapper[4979]: I1204 12:58:54.702150 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-f77rl" event={"ID":"049f0bfd-88d6-40b1-8be8-7d321de0c0dd","Type":"ContainerStarted","Data":"2c912e4fdd3eb40caa5b3658ca64b26a2be5f72d8b05b503cce04d93a3f18acd"} Dec 04 12:58:56 crc kubenswrapper[4979]: I1204 12:58:56.204729 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:58:56 crc kubenswrapper[4979]: E1204 12:58:56.205290 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:58:56 crc kubenswrapper[4979]: I1204 12:58:56.717821 4979 generic.go:334] "Generic (PLEG): container finished" podID="049f0bfd-88d6-40b1-8be8-7d321de0c0dd" containerID="475a0821eeb4bc52f9dc3d67428ad5c01b618195c83686a1a207e49a24035164" exitCode=0 Dec 04 12:58:56 crc kubenswrapper[4979]: I1204 12:58:56.717880 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-f77rl" event={"ID":"049f0bfd-88d6-40b1-8be8-7d321de0c0dd","Type":"ContainerDied","Data":"475a0821eeb4bc52f9dc3d67428ad5c01b618195c83686a1a207e49a24035164"} Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.031032 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.056039 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-crc-storage\") pod \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.056080 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22kgn\" (UniqueName: \"kubernetes.io/projected/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-kube-api-access-22kgn\") pod \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.056171 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-node-mnt\") pod \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\" (UID: \"049f0bfd-88d6-40b1-8be8-7d321de0c0dd\") " Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.056506 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "049f0bfd-88d6-40b1-8be8-7d321de0c0dd" (UID: "049f0bfd-88d6-40b1-8be8-7d321de0c0dd"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.061667 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-kube-api-access-22kgn" (OuterVolumeSpecName: "kube-api-access-22kgn") pod "049f0bfd-88d6-40b1-8be8-7d321de0c0dd" (UID: "049f0bfd-88d6-40b1-8be8-7d321de0c0dd"). InnerVolumeSpecName "kube-api-access-22kgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.077555 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "049f0bfd-88d6-40b1-8be8-7d321de0c0dd" (UID: "049f0bfd-88d6-40b1-8be8-7d321de0c0dd"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.157897 4979 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.157933 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22kgn\" (UniqueName: \"kubernetes.io/projected/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-kube-api-access-22kgn\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.157942 4979 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/049f0bfd-88d6-40b1-8be8-7d321de0c0dd-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.734091 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-f77rl" event={"ID":"049f0bfd-88d6-40b1-8be8-7d321de0c0dd","Type":"ContainerDied","Data":"2c912e4fdd3eb40caa5b3658ca64b26a2be5f72d8b05b503cce04d93a3f18acd"} Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.734440 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c912e4fdd3eb40caa5b3658ca64b26a2be5f72d8b05b503cce04d93a3f18acd" Dec 04 12:58:58 crc kubenswrapper[4979]: I1204 12:58:58.734260 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-f77rl" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.354986 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["crc-storage/crc-storage-crc-f77rl"] Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.360701 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["crc-storage/crc-storage-crc-f77rl"] Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.489617 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["crc-storage/crc-storage-crc-ksgw2"] Dec 04 12:59:00 crc kubenswrapper[4979]: E1204 12:59:00.489963 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="049f0bfd-88d6-40b1-8be8-7d321de0c0dd" containerName="storage" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.489984 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="049f0bfd-88d6-40b1-8be8-7d321de0c0dd" containerName="storage" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.490118 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="049f0bfd-88d6-40b1-8be8-7d321de0c0dd" containerName="storage" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.490628 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.492413 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"crc-storage" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.492595 4979 reflector.go:368] Caches populated for *v1.Secret from object-"crc-storage"/"crc-storage-dockercfg-tm2n7" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.492830 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"kube-root-ca.crt" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.493604 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"crc-storage"/"openshift-service-ca.crt" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.501186 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-ksgw2"] Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.586338 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/864739c9-728f-48ac-9ad7-6f08bde9a5ce-node-mnt\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.586440 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/864739c9-728f-48ac-9ad7-6f08bde9a5ce-crc-storage\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.586528 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtzvj\" (UniqueName: \"kubernetes.io/projected/864739c9-728f-48ac-9ad7-6f08bde9a5ce-kube-api-access-vtzvj\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.687740 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtzvj\" (UniqueName: \"kubernetes.io/projected/864739c9-728f-48ac-9ad7-6f08bde9a5ce-kube-api-access-vtzvj\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.687816 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/864739c9-728f-48ac-9ad7-6f08bde9a5ce-node-mnt\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.687945 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/864739c9-728f-48ac-9ad7-6f08bde9a5ce-crc-storage\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.688127 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/864739c9-728f-48ac-9ad7-6f08bde9a5ce-node-mnt\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.688827 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/864739c9-728f-48ac-9ad7-6f08bde9a5ce-crc-storage\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.707108 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtzvj\" (UniqueName: \"kubernetes.io/projected/864739c9-728f-48ac-9ad7-6f08bde9a5ce-kube-api-access-vtzvj\") pod \"crc-storage-crc-ksgw2\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:00 crc kubenswrapper[4979]: I1204 12:59:00.806415 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:01 crc kubenswrapper[4979]: I1204 12:59:01.220539 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["crc-storage/crc-storage-crc-ksgw2"] Dec 04 12:59:01 crc kubenswrapper[4979]: W1204 12:59:01.225234 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod864739c9_728f_48ac_9ad7_6f08bde9a5ce.slice/crio-0823f6c99f0f1439d0ef6ecd62ffeddf66220ac849c8b1244b4f2decdacc71bd WatchSource:0}: Error finding container 0823f6c99f0f1439d0ef6ecd62ffeddf66220ac849c8b1244b4f2decdacc71bd: Status 404 returned error can't find the container with id 0823f6c99f0f1439d0ef6ecd62ffeddf66220ac849c8b1244b4f2decdacc71bd Dec 04 12:59:01 crc kubenswrapper[4979]: I1204 12:59:01.756714 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-ksgw2" event={"ID":"864739c9-728f-48ac-9ad7-6f08bde9a5ce","Type":"ContainerStarted","Data":"0823f6c99f0f1439d0ef6ecd62ffeddf66220ac849c8b1244b4f2decdacc71bd"} Dec 04 12:59:02 crc kubenswrapper[4979]: I1204 12:59:02.211422 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="049f0bfd-88d6-40b1-8be8-7d321de0c0dd" path="/var/lib/kubelet/pods/049f0bfd-88d6-40b1-8be8-7d321de0c0dd/volumes" Dec 04 12:59:02 crc kubenswrapper[4979]: I1204 12:59:02.765907 4979 generic.go:334] "Generic (PLEG): container finished" podID="864739c9-728f-48ac-9ad7-6f08bde9a5ce" containerID="d0162c07ddda95879a4cc3a07c64ddb34d6d7f74e309bf44a7782aae5866978d" exitCode=0 Dec 04 12:59:02 crc kubenswrapper[4979]: I1204 12:59:02.765949 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-ksgw2" event={"ID":"864739c9-728f-48ac-9ad7-6f08bde9a5ce","Type":"ContainerDied","Data":"d0162c07ddda95879a4cc3a07c64ddb34d6d7f74e309bf44a7782aae5866978d"} Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.039384 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.134174 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/864739c9-728f-48ac-9ad7-6f08bde9a5ce-node-mnt\") pod \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.134256 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtzvj\" (UniqueName: \"kubernetes.io/projected/864739c9-728f-48ac-9ad7-6f08bde9a5ce-kube-api-access-vtzvj\") pod \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.134286 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/864739c9-728f-48ac-9ad7-6f08bde9a5ce-crc-storage\") pod \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\" (UID: \"864739c9-728f-48ac-9ad7-6f08bde9a5ce\") " Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.134274 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/864739c9-728f-48ac-9ad7-6f08bde9a5ce-node-mnt" (OuterVolumeSpecName: "node-mnt") pod "864739c9-728f-48ac-9ad7-6f08bde9a5ce" (UID: "864739c9-728f-48ac-9ad7-6f08bde9a5ce"). InnerVolumeSpecName "node-mnt". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.134544 4979 reconciler_common.go:293] "Volume detached for volume \"node-mnt\" (UniqueName: \"kubernetes.io/host-path/864739c9-728f-48ac-9ad7-6f08bde9a5ce-node-mnt\") on node \"crc\" DevicePath \"\"" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.139105 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/864739c9-728f-48ac-9ad7-6f08bde9a5ce-kube-api-access-vtzvj" (OuterVolumeSpecName: "kube-api-access-vtzvj") pod "864739c9-728f-48ac-9ad7-6f08bde9a5ce" (UID: "864739c9-728f-48ac-9ad7-6f08bde9a5ce"). InnerVolumeSpecName "kube-api-access-vtzvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.151047 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/864739c9-728f-48ac-9ad7-6f08bde9a5ce-crc-storage" (OuterVolumeSpecName: "crc-storage") pod "864739c9-728f-48ac-9ad7-6f08bde9a5ce" (UID: "864739c9-728f-48ac-9ad7-6f08bde9a5ce"). InnerVolumeSpecName "crc-storage". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.235509 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtzvj\" (UniqueName: \"kubernetes.io/projected/864739c9-728f-48ac-9ad7-6f08bde9a5ce-kube-api-access-vtzvj\") on node \"crc\" DevicePath \"\"" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.235540 4979 reconciler_common.go:293] "Volume detached for volume \"crc-storage\" (UniqueName: \"kubernetes.io/configmap/864739c9-728f-48ac-9ad7-6f08bde9a5ce-crc-storage\") on node \"crc\" DevicePath \"\"" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.781161 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="crc-storage/crc-storage-crc-ksgw2" event={"ID":"864739c9-728f-48ac-9ad7-6f08bde9a5ce","Type":"ContainerDied","Data":"0823f6c99f0f1439d0ef6ecd62ffeddf66220ac849c8b1244b4f2decdacc71bd"} Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.781222 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="crc-storage/crc-storage-crc-ksgw2" Dec 04 12:59:04 crc kubenswrapper[4979]: I1204 12:59:04.781240 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0823f6c99f0f1439d0ef6ecd62ffeddf66220ac849c8b1244b4f2decdacc71bd" Dec 04 12:59:05 crc kubenswrapper[4979]: I1204 12:59:05.854452 4979 scope.go:117] "RemoveContainer" containerID="e0f0996b0e0d34d127adb5109b2047117fcda81c7172b75f4198047276c964fe" Dec 04 12:59:10 crc kubenswrapper[4979]: I1204 12:59:10.199084 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:59:10 crc kubenswrapper[4979]: E1204 12:59:10.199866 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:59:24 crc kubenswrapper[4979]: I1204 12:59:24.199641 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:59:24 crc kubenswrapper[4979]: E1204 12:59:24.200431 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 12:59:38 crc kubenswrapper[4979]: I1204 12:59:38.198807 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 12:59:39 crc kubenswrapper[4979]: I1204 12:59:39.042080 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"2cb5c9544ef934945129415b00a3831fb523499e2744b67a856f00423a013073"} Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.180554 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm"] Dec 04 13:00:00 crc kubenswrapper[4979]: E1204 13:00:00.182528 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="864739c9-728f-48ac-9ad7-6f08bde9a5ce" containerName="storage" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.182559 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="864739c9-728f-48ac-9ad7-6f08bde9a5ce" containerName="storage" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.182813 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="864739c9-728f-48ac-9ad7-6f08bde9a5ce" containerName="storage" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.183545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.186984 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.187446 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.187702 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm"] Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.356146 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nf5g\" (UniqueName: \"kubernetes.io/projected/154bfa25-b814-486a-808f-c61bcba05297-kube-api-access-8nf5g\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.356269 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/154bfa25-b814-486a-808f-c61bcba05297-secret-volume\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.356339 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/154bfa25-b814-486a-808f-c61bcba05297-config-volume\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.457926 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/154bfa25-b814-486a-808f-c61bcba05297-secret-volume\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.457998 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/154bfa25-b814-486a-808f-c61bcba05297-config-volume\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.458069 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nf5g\" (UniqueName: \"kubernetes.io/projected/154bfa25-b814-486a-808f-c61bcba05297-kube-api-access-8nf5g\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.459127 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/154bfa25-b814-486a-808f-c61bcba05297-config-volume\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.464088 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/154bfa25-b814-486a-808f-c61bcba05297-secret-volume\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.474479 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nf5g\" (UniqueName: \"kubernetes.io/projected/154bfa25-b814-486a-808f-c61bcba05297-kube-api-access-8nf5g\") pod \"collect-profiles-29414220-n2mmm\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.509458 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:00 crc kubenswrapper[4979]: I1204 13:00:00.923748 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm"] Dec 04 13:00:01 crc kubenswrapper[4979]: I1204 13:00:01.193532 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" event={"ID":"154bfa25-b814-486a-808f-c61bcba05297","Type":"ContainerStarted","Data":"6865f4cd9a4cd0065a5531a535a8e964e4b2383bc7cc2a9ddff8b6fc73785ecc"} Dec 04 13:00:01 crc kubenswrapper[4979]: I1204 13:00:01.194535 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" event={"ID":"154bfa25-b814-486a-808f-c61bcba05297","Type":"ContainerStarted","Data":"acc034b56ee45a0ed9657f6b1c3f920b4f583e514c13b5ff2560475d9b75cae6"} Dec 04 13:00:01 crc kubenswrapper[4979]: I1204 13:00:01.218340 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" podStartSLOduration=1.218308999 podStartE2EDuration="1.218308999s" podCreationTimestamp="2025-12-04 13:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:00:01.212191683 +0000 UTC m=+4625.486487487" watchObservedRunningTime="2025-12-04 13:00:01.218308999 +0000 UTC m=+4625.492604803" Dec 04 13:00:02 crc kubenswrapper[4979]: I1204 13:00:02.202468 4979 generic.go:334] "Generic (PLEG): container finished" podID="154bfa25-b814-486a-808f-c61bcba05297" containerID="6865f4cd9a4cd0065a5531a535a8e964e4b2383bc7cc2a9ddff8b6fc73785ecc" exitCode=0 Dec 04 13:00:02 crc kubenswrapper[4979]: I1204 13:00:02.207009 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" event={"ID":"154bfa25-b814-486a-808f-c61bcba05297","Type":"ContainerDied","Data":"6865f4cd9a4cd0065a5531a535a8e964e4b2383bc7cc2a9ddff8b6fc73785ecc"} Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.634212 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.803126 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/154bfa25-b814-486a-808f-c61bcba05297-config-volume\") pod \"154bfa25-b814-486a-808f-c61bcba05297\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.803178 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/154bfa25-b814-486a-808f-c61bcba05297-secret-volume\") pod \"154bfa25-b814-486a-808f-c61bcba05297\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.803272 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nf5g\" (UniqueName: \"kubernetes.io/projected/154bfa25-b814-486a-808f-c61bcba05297-kube-api-access-8nf5g\") pod \"154bfa25-b814-486a-808f-c61bcba05297\" (UID: \"154bfa25-b814-486a-808f-c61bcba05297\") " Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.803915 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/154bfa25-b814-486a-808f-c61bcba05297-config-volume" (OuterVolumeSpecName: "config-volume") pod "154bfa25-b814-486a-808f-c61bcba05297" (UID: "154bfa25-b814-486a-808f-c61bcba05297"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.809227 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/154bfa25-b814-486a-808f-c61bcba05297-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "154bfa25-b814-486a-808f-c61bcba05297" (UID: "154bfa25-b814-486a-808f-c61bcba05297"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.809721 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154bfa25-b814-486a-808f-c61bcba05297-kube-api-access-8nf5g" (OuterVolumeSpecName: "kube-api-access-8nf5g") pod "154bfa25-b814-486a-808f-c61bcba05297" (UID: "154bfa25-b814-486a-808f-c61bcba05297"). InnerVolumeSpecName "kube-api-access-8nf5g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.905088 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nf5g\" (UniqueName: \"kubernetes.io/projected/154bfa25-b814-486a-808f-c61bcba05297-kube-api-access-8nf5g\") on node \"crc\" DevicePath \"\"" Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.905138 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/154bfa25-b814-486a-808f-c61bcba05297-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:00:03 crc kubenswrapper[4979]: I1204 13:00:03.905150 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/154bfa25-b814-486a-808f-c61bcba05297-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:00:04 crc kubenswrapper[4979]: I1204 13:00:04.219557 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" event={"ID":"154bfa25-b814-486a-808f-c61bcba05297","Type":"ContainerDied","Data":"acc034b56ee45a0ed9657f6b1c3f920b4f583e514c13b5ff2560475d9b75cae6"} Dec 04 13:00:04 crc kubenswrapper[4979]: I1204 13:00:04.219596 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="acc034b56ee45a0ed9657f6b1c3f920b4f583e514c13b5ff2560475d9b75cae6" Dec 04 13:00:04 crc kubenswrapper[4979]: I1204 13:00:04.219637 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm" Dec 04 13:00:04 crc kubenswrapper[4979]: I1204 13:00:04.705361 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm"] Dec 04 13:00:04 crc kubenswrapper[4979]: I1204 13:00:04.709836 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414175-w8gsm"] Dec 04 13:00:06 crc kubenswrapper[4979]: I1204 13:00:06.208335 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="164112b0-ceef-4b99-8de0-fdd458ab6354" path="/var/lib/kubelet/pods/164112b0-ceef-4b99-8de0-fdd458ab6354/volumes" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.128069 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x5krh"] Dec 04 13:00:49 crc kubenswrapper[4979]: E1204 13:00:49.129144 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154bfa25-b814-486a-808f-c61bcba05297" containerName="collect-profiles" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.129158 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="154bfa25-b814-486a-808f-c61bcba05297" containerName="collect-profiles" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.129340 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="154bfa25-b814-486a-808f-c61bcba05297" containerName="collect-profiles" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.130488 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.140782 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5krh"] Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.188680 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-utilities\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.188762 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsw4j\" (UniqueName: \"kubernetes.io/projected/e3887d27-b761-45d8-a156-39afaf3fb982-kube-api-access-lsw4j\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.188803 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-catalog-content\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.293643 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsw4j\" (UniqueName: \"kubernetes.io/projected/e3887d27-b761-45d8-a156-39afaf3fb982-kube-api-access-lsw4j\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.294072 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-catalog-content\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.294248 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-utilities\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.294828 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-utilities\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.295120 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-catalog-content\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.571905 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsw4j\" (UniqueName: \"kubernetes.io/projected/e3887d27-b761-45d8-a156-39afaf3fb982-kube-api-access-lsw4j\") pod \"redhat-marketplace-x5krh\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:49 crc kubenswrapper[4979]: I1204 13:00:49.762833 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:50 crc kubenswrapper[4979]: I1204 13:00:50.177913 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5krh"] Dec 04 13:00:50 crc kubenswrapper[4979]: I1204 13:00:50.936778 4979 generic.go:334] "Generic (PLEG): container finished" podID="e3887d27-b761-45d8-a156-39afaf3fb982" containerID="2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1" exitCode=0 Dec 04 13:00:50 crc kubenswrapper[4979]: I1204 13:00:50.937039 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5krh" event={"ID":"e3887d27-b761-45d8-a156-39afaf3fb982","Type":"ContainerDied","Data":"2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1"} Dec 04 13:00:50 crc kubenswrapper[4979]: I1204 13:00:50.937071 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5krh" event={"ID":"e3887d27-b761-45d8-a156-39afaf3fb982","Type":"ContainerStarted","Data":"f862e2c733d8921125e4c86767f9be15fc4e6d7dd9610d08dd9e9d48521513d0"} Dec 04 13:00:50 crc kubenswrapper[4979]: I1204 13:00:50.938945 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:00:51 crc kubenswrapper[4979]: I1204 13:00:51.947386 4979 generic.go:334] "Generic (PLEG): container finished" podID="e3887d27-b761-45d8-a156-39afaf3fb982" containerID="0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57" exitCode=0 Dec 04 13:00:51 crc kubenswrapper[4979]: I1204 13:00:51.947484 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5krh" event={"ID":"e3887d27-b761-45d8-a156-39afaf3fb982","Type":"ContainerDied","Data":"0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57"} Dec 04 13:00:52 crc kubenswrapper[4979]: I1204 13:00:52.957820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5krh" event={"ID":"e3887d27-b761-45d8-a156-39afaf3fb982","Type":"ContainerStarted","Data":"9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f"} Dec 04 13:00:52 crc kubenswrapper[4979]: I1204 13:00:52.978781 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x5krh" podStartSLOduration=2.446846489 podStartE2EDuration="3.978759543s" podCreationTimestamp="2025-12-04 13:00:49 +0000 UTC" firstStartedPulling="2025-12-04 13:00:50.938696029 +0000 UTC m=+4675.212991833" lastFinishedPulling="2025-12-04 13:00:52.470609083 +0000 UTC m=+4676.744904887" observedRunningTime="2025-12-04 13:00:52.97787639 +0000 UTC m=+4677.252172204" watchObservedRunningTime="2025-12-04 13:00:52.978759543 +0000 UTC m=+4677.253055347" Dec 04 13:00:59 crc kubenswrapper[4979]: I1204 13:00:59.764067 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:59 crc kubenswrapper[4979]: I1204 13:00:59.764476 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:00:59 crc kubenswrapper[4979]: I1204 13:00:59.805010 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:01:00 crc kubenswrapper[4979]: I1204 13:01:00.047378 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:01:00 crc kubenswrapper[4979]: I1204 13:01:00.092217 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5krh"] Dec 04 13:01:02 crc kubenswrapper[4979]: I1204 13:01:02.014987 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x5krh" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="registry-server" containerID="cri-o://9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f" gracePeriod=2 Dec 04 13:01:02 crc kubenswrapper[4979]: I1204 13:01:02.880277 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.023593 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5krh" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.023597 4979 generic.go:334] "Generic (PLEG): container finished" podID="e3887d27-b761-45d8-a156-39afaf3fb982" containerID="9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f" exitCode=0 Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.023877 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5krh" event={"ID":"e3887d27-b761-45d8-a156-39afaf3fb982","Type":"ContainerDied","Data":"9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f"} Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.023968 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5krh" event={"ID":"e3887d27-b761-45d8-a156-39afaf3fb982","Type":"ContainerDied","Data":"f862e2c733d8921125e4c86767f9be15fc4e6d7dd9610d08dd9e9d48521513d0"} Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.024018 4979 scope.go:117] "RemoveContainer" containerID="9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.048016 4979 scope.go:117] "RemoveContainer" containerID="0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.071486 4979 scope.go:117] "RemoveContainer" containerID="2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.080178 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-utilities\") pod \"e3887d27-b761-45d8-a156-39afaf3fb982\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.080409 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-catalog-content\") pod \"e3887d27-b761-45d8-a156-39afaf3fb982\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.080477 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsw4j\" (UniqueName: \"kubernetes.io/projected/e3887d27-b761-45d8-a156-39afaf3fb982-kube-api-access-lsw4j\") pod \"e3887d27-b761-45d8-a156-39afaf3fb982\" (UID: \"e3887d27-b761-45d8-a156-39afaf3fb982\") " Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.081141 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-utilities" (OuterVolumeSpecName: "utilities") pod "e3887d27-b761-45d8-a156-39afaf3fb982" (UID: "e3887d27-b761-45d8-a156-39afaf3fb982"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.099657 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3887d27-b761-45d8-a156-39afaf3fb982-kube-api-access-lsw4j" (OuterVolumeSpecName: "kube-api-access-lsw4j") pod "e3887d27-b761-45d8-a156-39afaf3fb982" (UID: "e3887d27-b761-45d8-a156-39afaf3fb982"). InnerVolumeSpecName "kube-api-access-lsw4j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.103174 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e3887d27-b761-45d8-a156-39afaf3fb982" (UID: "e3887d27-b761-45d8-a156-39afaf3fb982"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.129411 4979 scope.go:117] "RemoveContainer" containerID="9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f" Dec 04 13:01:03 crc kubenswrapper[4979]: E1204 13:01:03.129808 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f\": container with ID starting with 9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f not found: ID does not exist" containerID="9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.129848 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f"} err="failed to get container status \"9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f\": rpc error: code = NotFound desc = could not find container \"9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f\": container with ID starting with 9a3ac0e978131641d4f0e1819716b434597facbbd41da6bc065a7800d58f6e2f not found: ID does not exist" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.129872 4979 scope.go:117] "RemoveContainer" containerID="0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57" Dec 04 13:01:03 crc kubenswrapper[4979]: E1204 13:01:03.130129 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57\": container with ID starting with 0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57 not found: ID does not exist" containerID="0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.130167 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57"} err="failed to get container status \"0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57\": rpc error: code = NotFound desc = could not find container \"0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57\": container with ID starting with 0122d74d62a9990eba498b946047db18b406ca3681b23a9814028806297d9b57 not found: ID does not exist" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.130193 4979 scope.go:117] "RemoveContainer" containerID="2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1" Dec 04 13:01:03 crc kubenswrapper[4979]: E1204 13:01:03.130673 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1\": container with ID starting with 2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1 not found: ID does not exist" containerID="2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.130697 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1"} err="failed to get container status \"2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1\": rpc error: code = NotFound desc = could not find container \"2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1\": container with ID starting with 2b23e3413c5f9d39b38e177ac6629c33daf40252a05eeb89f1df4bc9fb6d5ff1 not found: ID does not exist" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.182082 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.182183 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsw4j\" (UniqueName: \"kubernetes.io/projected/e3887d27-b761-45d8-a156-39afaf3fb982-kube-api-access-lsw4j\") on node \"crc\" DevicePath \"\"" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.182228 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e3887d27-b761-45d8-a156-39afaf3fb982-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.354886 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5krh"] Dec 04 13:01:03 crc kubenswrapper[4979]: I1204 13:01:03.360629 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5krh"] Dec 04 13:01:04 crc kubenswrapper[4979]: I1204 13:01:04.208677 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" path="/var/lib/kubelet/pods/e3887d27-b761-45d8-a156-39afaf3fb982/volumes" Dec 04 13:01:05 crc kubenswrapper[4979]: I1204 13:01:05.926681 4979 scope.go:117] "RemoveContainer" containerID="594a1a3de87a48be0a9ccf3c5c5b9e4706868d1fc7c08f67a295e507a14d6367" Dec 04 13:01:58 crc kubenswrapper[4979]: I1204 13:01:58.040921 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:01:58 crc kubenswrapper[4979]: I1204 13:01:58.041485 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.632953 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-847qc"] Dec 04 13:02:08 crc kubenswrapper[4979]: E1204 13:02:08.633718 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="registry-server" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.633731 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="registry-server" Dec 04 13:02:08 crc kubenswrapper[4979]: E1204 13:02:08.633750 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="extract-utilities" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.633756 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="extract-utilities" Dec 04 13:02:08 crc kubenswrapper[4979]: E1204 13:02:08.633774 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="extract-content" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.633780 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="extract-content" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.633905 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3887d27-b761-45d8-a156-39afaf3fb982" containerName="registry-server" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.634616 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.639506 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.639699 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.639807 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.639928 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.640097 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-lsr7c" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.644568 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-847qc"] Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.694789 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-config\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.695094 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.695265 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qkrh\" (UniqueName: \"kubernetes.io/projected/575217c8-729c-4491-8d04-076ed4184afc-kube-api-access-9qkrh\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.796783 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-config\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.796853 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.796905 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qkrh\" (UniqueName: \"kubernetes.io/projected/575217c8-729c-4491-8d04-076ed4184afc-kube-api-access-9qkrh\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.797937 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.797971 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-config\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.824465 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qkrh\" (UniqueName: \"kubernetes.io/projected/575217c8-729c-4491-8d04-076ed4184afc-kube-api-access-9qkrh\") pod \"dnsmasq-dns-5d7b5456f5-847qc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:08 crc kubenswrapper[4979]: I1204 13:02:08.950721 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.027673 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-7xlrw"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.029284 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.058718 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-7xlrw"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.102248 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-config\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.102392 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.102422 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppgs5\" (UniqueName: \"kubernetes.io/projected/8bf461b9-e97a-46ff-95c3-6f32f568e87f-kube-api-access-ppgs5\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.203363 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.203411 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppgs5\" (UniqueName: \"kubernetes.io/projected/8bf461b9-e97a-46ff-95c3-6f32f568e87f-kube-api-access-ppgs5\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.203452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-config\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.204236 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-dns-svc\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.204288 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-config\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.237618 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppgs5\" (UniqueName: \"kubernetes.io/projected/8bf461b9-e97a-46ff-95c3-6f32f568e87f-kube-api-access-ppgs5\") pod \"dnsmasq-dns-98ddfc8f-7xlrw\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.362518 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.577077 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-847qc"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.780125 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.781551 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.785398 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wbpgc" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.785399 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.786141 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.786414 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.786613 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.788777 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815464 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815508 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815557 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e848001-e03d-4db6-84f9-24fecba801e4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815598 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815664 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815685 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e848001-e03d-4db6-84f9-24fecba801e4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815707 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6vwf\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-kube-api-access-p6vwf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.815724 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.832469 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.840160 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.844368 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.854867 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-w8zjt" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.873085 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.900288 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-7xlrw"] Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916800 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916857 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916881 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e848001-e03d-4db6-84f9-24fecba801e4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916902 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af82841c-53fb-4e62-8fd3-fbf888e87045-kolla-config\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916934 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpmgl\" (UniqueName: \"kubernetes.io/projected/af82841c-53fb-4e62-8fd3-fbf888e87045-kube-api-access-tpmgl\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916954 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6vwf\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-kube-api-access-p6vwf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.916968 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.917031 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.917048 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.917065 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af82841c-53fb-4e62-8fd3-fbf888e87045-config-data\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.917096 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.917116 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e848001-e03d-4db6-84f9-24fecba801e4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.918770 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.920673 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.920913 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.921637 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.924750 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.930736 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e848001-e03d-4db6-84f9-24fecba801e4-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.934917 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.934977 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2be5edbcc6da3df1b51063102b8f2463830666ce99ffb0e08566e12b8abd0f29/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.938923 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6vwf\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-kube-api-access-p6vwf\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.940235 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e848001-e03d-4db6-84f9-24fecba801e4-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:09 crc kubenswrapper[4979]: I1204 13:02:09.987224 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " pod="openstack/rabbitmq-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.019070 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af82841c-53fb-4e62-8fd3-fbf888e87045-config-data\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.019180 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af82841c-53fb-4e62-8fd3-fbf888e87045-kolla-config\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.019200 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpmgl\" (UniqueName: \"kubernetes.io/projected/af82841c-53fb-4e62-8fd3-fbf888e87045-kube-api-access-tpmgl\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.020084 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/af82841c-53fb-4e62-8fd3-fbf888e87045-kolla-config\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.020603 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af82841c-53fb-4e62-8fd3-fbf888e87045-config-data\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.035269 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpmgl\" (UniqueName: \"kubernetes.io/projected/af82841c-53fb-4e62-8fd3-fbf888e87045-kube-api-access-tpmgl\") pod \"memcached-0\" (UID: \"af82841c-53fb-4e62-8fd3-fbf888e87045\") " pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.090021 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.094803 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.155939 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.157220 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.159653 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-zspgk" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.159795 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.160388 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.160803 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.170797 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.183964 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.223858 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4f9a655-3c5c-4053-bf72-1b99d3631889-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.223907 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-87b3ea17-b7bc-4842-9449-7a957934054e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b3ea17-b7bc-4842-9449-7a957934054e\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.223936 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4f9a655-3c5c-4053-bf72-1b99d3631889-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.223959 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.224069 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjwfc\" (UniqueName: \"kubernetes.io/projected/a4f9a655-3c5c-4053-bf72-1b99d3631889-kube-api-access-hjwfc\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.224107 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-config-data-default\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.224146 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a655-3c5c-4053-bf72-1b99d3631889-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.224185 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-kolla-config\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.225770 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.227361 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.229783 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.229867 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.229945 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.229866 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2gx5p" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.230010 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.244155 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325430 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325519 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjwfc\" (UniqueName: \"kubernetes.io/projected/a4f9a655-3c5c-4053-bf72-1b99d3631889-kube-api-access-hjwfc\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325540 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-config-data-default\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325563 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a655-3c5c-4053-bf72-1b99d3631889-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325584 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-kolla-config\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325604 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4f9a655-3c5c-4053-bf72-1b99d3631889-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325624 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-87b3ea17-b7bc-4842-9449-7a957934054e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b3ea17-b7bc-4842-9449-7a957934054e\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.325645 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4f9a655-3c5c-4053-bf72-1b99d3631889-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.326851 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a4f9a655-3c5c-4053-bf72-1b99d3631889-config-data-generated\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.328490 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-operator-scripts\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.329040 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-kolla-config\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.329528 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a4f9a655-3c5c-4053-bf72-1b99d3631889-config-data-default\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.330234 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4f9a655-3c5c-4053-bf72-1b99d3631889-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.330517 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a4f9a655-3c5c-4053-bf72-1b99d3631889-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.336207 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.336254 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-87b3ea17-b7bc-4842-9449-7a957934054e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b3ea17-b7bc-4842-9449-7a957934054e\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/56e3e78b600ff873714c39232c4311b85e2287c5d8d77a05d3927d2b8a6524ad/globalmount\"" pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.347379 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjwfc\" (UniqueName: \"kubernetes.io/projected/a4f9a655-3c5c-4053-bf72-1b99d3631889-kube-api-access-hjwfc\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.370175 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-87b3ea17-b7bc-4842-9449-7a957934054e\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87b3ea17-b7bc-4842-9449-7a957934054e\") pod \"openstack-galera-0\" (UID: \"a4f9a655-3c5c-4053-bf72-1b99d3631889\") " pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.401807 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: W1204 13:02:10.404897 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf82841c_53fb_4e62_8fd3_fbf888e87045.slice/crio-64beedc8ae8bb0df136b166dd1fb8b47a02c95dc4e2d0fe67bdaa51d9c9687d6 WatchSource:0}: Error finding container 64beedc8ae8bb0df136b166dd1fb8b47a02c95dc4e2d0fe67bdaa51d9c9687d6: Status 404 returned error can't find the container with id 64beedc8ae8bb0df136b166dd1fb8b47a02c95dc4e2d0fe67bdaa51d9c9687d6 Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.426694 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.426756 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.426924 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.427062 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.427409 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.427509 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9898\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-kube-api-access-s9898\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.427607 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.427671 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.427769 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.522987 4979 generic.go:334] "Generic (PLEG): container finished" podID="575217c8-729c-4491-8d04-076ed4184afc" containerID="558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2" exitCode=0 Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.523047 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" event={"ID":"575217c8-729c-4491-8d04-076ed4184afc","Type":"ContainerDied","Data":"558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2"} Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.523103 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" event={"ID":"575217c8-729c-4491-8d04-076ed4184afc","Type":"ContainerStarted","Data":"2376ee5772db2d3f62269eb7867e3d6f4f0ccdf6969787e28d4ab207e2f49303"} Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.524635 4979 generic.go:334] "Generic (PLEG): container finished" podID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerID="2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c" exitCode=0 Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.524731 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" event={"ID":"8bf461b9-e97a-46ff-95c3-6f32f568e87f","Type":"ContainerDied","Data":"2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c"} Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.524767 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" event={"ID":"8bf461b9-e97a-46ff-95c3-6f32f568e87f","Type":"ContainerStarted","Data":"bef695a232dd93976ddf75974b6d2a9a461fb07cd80c40dcf7eb626dbb25762f"} Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.526730 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530833 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530871 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9898\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-kube-api-access-s9898\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530899 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530920 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530941 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530959 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.530983 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.531008 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.531035 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.533068 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"af82841c-53fb-4e62-8fd3-fbf888e87045","Type":"ContainerStarted","Data":"87631857a27e502efe8be427b8b38fdd6a8476fe665caed800a35b53f044f532"} Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.533110 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"af82841c-53fb-4e62-8fd3-fbf888e87045","Type":"ContainerStarted","Data":"64beedc8ae8bb0df136b166dd1fb8b47a02c95dc4e2d0fe67bdaa51d9c9687d6"} Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.533474 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.533583 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.533891 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.534122 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.534781 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.535418 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.535448 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c04b0c08aa6505c08e21eaa4ffb138669ded1689a8c37464271f4824abcccd00/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.537101 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.537254 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.539632 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.544581 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.555358 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9898\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-kube-api-access-s9898\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.588602 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.588582668 podStartE2EDuration="1.588582668s" podCreationTimestamp="2025-12-04 13:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:10.58644819 +0000 UTC m=+4754.860744024" watchObservedRunningTime="2025-12-04 13:02:10.588582668 +0000 UTC m=+4754.862878482" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.590823 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: E1204 13:02:10.706084 4979 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 04 13:02:10 crc kubenswrapper[4979]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/575217c8-729c-4491-8d04-076ed4184afc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 04 13:02:10 crc kubenswrapper[4979]: > podSandboxID="2376ee5772db2d3f62269eb7867e3d6f4f0ccdf6969787e28d4ab207e2f49303" Dec 04 13:02:10 crc kubenswrapper[4979]: E1204 13:02:10.706262 4979 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 04 13:02:10 crc kubenswrapper[4979]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n8chc6h5bh56fh546hb7hc8h67h5bchffh577h697h5b5h5bdh59bhf6hf4h558hb5h578h595h5cchfbh644h59ch7fh654h547h587h5cbh5d5h8fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9qkrh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5d7b5456f5-847qc_openstack(575217c8-729c-4491-8d04-076ed4184afc): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/575217c8-729c-4491-8d04-076ed4184afc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 04 13:02:10 crc kubenswrapper[4979]: > logger="UnhandledError" Dec 04 13:02:10 crc kubenswrapper[4979]: E1204 13:02:10.707463 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/575217c8-729c-4491-8d04-076ed4184afc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" podUID="575217c8-729c-4491-8d04-076ed4184afc" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.847803 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.970388 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 04 13:02:10 crc kubenswrapper[4979]: I1204 13:02:10.998782 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.001228 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.005837 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-dnxwp" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.006045 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.006178 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.006278 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.008059 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140035 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140086 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-a0874a85-1850-477a-8308-b34eb3b907d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0874a85-1850-477a-8308-b34eb3b907d6\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140115 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140359 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140457 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140564 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wtqf2\" (UniqueName: \"kubernetes.io/projected/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-kube-api-access-wtqf2\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140632 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.140674 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243011 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243355 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-a0874a85-1850-477a-8308-b34eb3b907d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0874a85-1850-477a-8308-b34eb3b907d6\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243391 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243421 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243447 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243494 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wtqf2\" (UniqueName: \"kubernetes.io/projected/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-kube-api-access-wtqf2\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243567 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.243989 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.244365 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.244872 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.245550 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.247044 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.247076 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-a0874a85-1850-477a-8308-b34eb3b907d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0874a85-1850-477a-8308-b34eb3b907d6\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/434d6155337d19294719ada2e4cff09fc93bd9988ac4e72a76996cadabd5a1aa/globalmount\"" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.249504 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.250016 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.262197 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wtqf2\" (UniqueName: \"kubernetes.io/projected/f6a16c64-130c-4c9a-988b-a3fa304c4d1e-kube-api-access-wtqf2\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.275982 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-a0874a85-1850-477a-8308-b34eb3b907d6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-a0874a85-1850-477a-8308-b34eb3b907d6\") pod \"openstack-cell1-galera-0\" (UID: \"f6a16c64-130c-4c9a-988b-a3fa304c4d1e\") " pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.293526 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:02:11 crc kubenswrapper[4979]: W1204 13:02:11.375930 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod673d7a64_3fbe_4fd8_9dae_d7b2a3edd0c2.slice/crio-93cc9b08fdf7e1cd6d407a8be70d3473c0c88ee1161d9e358865e3272d7f0db3 WatchSource:0}: Error finding container 93cc9b08fdf7e1cd6d407a8be70d3473c0c88ee1161d9e358865e3272d7f0db3: Status 404 returned error can't find the container with id 93cc9b08fdf7e1cd6d407a8be70d3473c0c88ee1161d9e358865e3272d7f0db3 Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.464362 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.548662 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a4f9a655-3c5c-4053-bf72-1b99d3631889","Type":"ContainerStarted","Data":"365f8f3f774f40aa0b7c156b197c9eed418c952e7b47dd0b942bc035e87749d7"} Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.554799 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e848001-e03d-4db6-84f9-24fecba801e4","Type":"ContainerStarted","Data":"cb885fc6a813660644d1a8caa6d508a64eef44e0d5f8c2b57397253da16d4278"} Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.558030 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2","Type":"ContainerStarted","Data":"93cc9b08fdf7e1cd6d407a8be70d3473c0c88ee1161d9e358865e3272d7f0db3"} Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.562535 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" event={"ID":"8bf461b9-e97a-46ff-95c3-6f32f568e87f","Type":"ContainerStarted","Data":"2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238"} Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.592570 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" podStartSLOduration=3.592545394 podStartE2EDuration="3.592545394s" podCreationTimestamp="2025-12-04 13:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:11.583518299 +0000 UTC m=+4755.857814153" watchObservedRunningTime="2025-12-04 13:02:11.592545394 +0000 UTC m=+4755.866841218" Dec 04 13:02:11 crc kubenswrapper[4979]: I1204 13:02:11.909831 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 04 13:02:11 crc kubenswrapper[4979]: W1204 13:02:11.912253 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6a16c64_130c_4c9a_988b_a3fa304c4d1e.slice/crio-44223bf5f3234a09683c54e573bf27de7bce0d516fbe576cfd05679e9f8c9220 WatchSource:0}: Error finding container 44223bf5f3234a09683c54e573bf27de7bce0d516fbe576cfd05679e9f8c9220: Status 404 returned error can't find the container with id 44223bf5f3234a09683c54e573bf27de7bce0d516fbe576cfd05679e9f8c9220 Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.570258 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f6a16c64-130c-4c9a-988b-a3fa304c4d1e","Type":"ContainerStarted","Data":"be89972317ab701803e070e75c994d3f33edd1b4871236c91f5f6186c448574a"} Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.570706 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f6a16c64-130c-4c9a-988b-a3fa304c4d1e","Type":"ContainerStarted","Data":"44223bf5f3234a09683c54e573bf27de7bce0d516fbe576cfd05679e9f8c9220"} Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.571741 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2","Type":"ContainerStarted","Data":"db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6"} Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.577745 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" event={"ID":"575217c8-729c-4491-8d04-076ed4184afc","Type":"ContainerStarted","Data":"e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826"} Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.578208 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.582486 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a4f9a655-3c5c-4053-bf72-1b99d3631889","Type":"ContainerStarted","Data":"c2f4f30b561cbc4f8a40f0d5d4fe72d6c3735bc824b041617a38f9a3439a21d1"} Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.585473 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e848001-e03d-4db6-84f9-24fecba801e4","Type":"ContainerStarted","Data":"ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4"} Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.585809 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:12 crc kubenswrapper[4979]: I1204 13:02:12.654289 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" podStartSLOduration=4.654273087 podStartE2EDuration="4.654273087s" podCreationTimestamp="2025-12-04 13:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:12.649264362 +0000 UTC m=+4756.923560166" watchObservedRunningTime="2025-12-04 13:02:12.654273087 +0000 UTC m=+4756.928568891" Dec 04 13:02:15 crc kubenswrapper[4979]: I1204 13:02:15.097155 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 04 13:02:15 crc kubenswrapper[4979]: I1204 13:02:15.607480 4979 generic.go:334] "Generic (PLEG): container finished" podID="a4f9a655-3c5c-4053-bf72-1b99d3631889" containerID="c2f4f30b561cbc4f8a40f0d5d4fe72d6c3735bc824b041617a38f9a3439a21d1" exitCode=0 Dec 04 13:02:15 crc kubenswrapper[4979]: I1204 13:02:15.607529 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a4f9a655-3c5c-4053-bf72-1b99d3631889","Type":"ContainerDied","Data":"c2f4f30b561cbc4f8a40f0d5d4fe72d6c3735bc824b041617a38f9a3439a21d1"} Dec 04 13:02:16 crc kubenswrapper[4979]: I1204 13:02:16.616867 4979 generic.go:334] "Generic (PLEG): container finished" podID="f6a16c64-130c-4c9a-988b-a3fa304c4d1e" containerID="be89972317ab701803e070e75c994d3f33edd1b4871236c91f5f6186c448574a" exitCode=0 Dec 04 13:02:16 crc kubenswrapper[4979]: I1204 13:02:16.616947 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f6a16c64-130c-4c9a-988b-a3fa304c4d1e","Type":"ContainerDied","Data":"be89972317ab701803e070e75c994d3f33edd1b4871236c91f5f6186c448574a"} Dec 04 13:02:16 crc kubenswrapper[4979]: I1204 13:02:16.620354 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"a4f9a655-3c5c-4053-bf72-1b99d3631889","Type":"ContainerStarted","Data":"3f8be7b8dcc8e6fc84df24f8a900e202d79b62bf5c753043de7690fb70a45b34"} Dec 04 13:02:16 crc kubenswrapper[4979]: I1204 13:02:16.675063 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=7.675035565 podStartE2EDuration="7.675035565s" podCreationTimestamp="2025-12-04 13:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:16.668016075 +0000 UTC m=+4760.942311889" watchObservedRunningTime="2025-12-04 13:02:16.675035565 +0000 UTC m=+4760.949331409" Dec 04 13:02:17 crc kubenswrapper[4979]: I1204 13:02:17.630279 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"f6a16c64-130c-4c9a-988b-a3fa304c4d1e","Type":"ContainerStarted","Data":"8625f64890568401dcbaabf0df4f5940704c5767bc8f616e17e59a5d97e9edbe"} Dec 04 13:02:17 crc kubenswrapper[4979]: I1204 13:02:17.653508 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.653489139 podStartE2EDuration="8.653489139s" podCreationTimestamp="2025-12-04 13:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:17.648327678 +0000 UTC m=+4761.922623492" watchObservedRunningTime="2025-12-04 13:02:17.653489139 +0000 UTC m=+4761.927784943" Dec 04 13:02:18 crc kubenswrapper[4979]: I1204 13:02:18.953061 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:19 crc kubenswrapper[4979]: I1204 13:02:19.364954 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:02:19 crc kubenswrapper[4979]: I1204 13:02:19.422868 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-847qc"] Dec 04 13:02:19 crc kubenswrapper[4979]: I1204 13:02:19.644486 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" podUID="575217c8-729c-4491-8d04-076ed4184afc" containerName="dnsmasq-dns" containerID="cri-o://e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826" gracePeriod=10 Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.187207 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.280602 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc\") pod \"575217c8-729c-4491-8d04-076ed4184afc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.280643 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-config\") pod \"575217c8-729c-4491-8d04-076ed4184afc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.280731 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qkrh\" (UniqueName: \"kubernetes.io/projected/575217c8-729c-4491-8d04-076ed4184afc-kube-api-access-9qkrh\") pod \"575217c8-729c-4491-8d04-076ed4184afc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.297948 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/575217c8-729c-4491-8d04-076ed4184afc-kube-api-access-9qkrh" (OuterVolumeSpecName: "kube-api-access-9qkrh") pod "575217c8-729c-4491-8d04-076ed4184afc" (UID: "575217c8-729c-4491-8d04-076ed4184afc"). InnerVolumeSpecName "kube-api-access-9qkrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:02:20 crc kubenswrapper[4979]: E1204 13:02:20.315233 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc podName:575217c8-729c-4491-8d04-076ed4184afc nodeName:}" failed. No retries permitted until 2025-12-04 13:02:20.815201444 +0000 UTC m=+4765.089497248 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "dns-svc" (UniqueName: "kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc") pod "575217c8-729c-4491-8d04-076ed4184afc" (UID: "575217c8-729c-4491-8d04-076ed4184afc") : error deleting /var/lib/kubelet/pods/575217c8-729c-4491-8d04-076ed4184afc/volume-subpaths: remove /var/lib/kubelet/pods/575217c8-729c-4491-8d04-076ed4184afc/volume-subpaths: no such file or directory Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.315482 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-config" (OuterVolumeSpecName: "config") pod "575217c8-729c-4491-8d04-076ed4184afc" (UID: "575217c8-729c-4491-8d04-076ed4184afc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.382276 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.382353 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qkrh\" (UniqueName: \"kubernetes.io/projected/575217c8-729c-4491-8d04-076ed4184afc-kube-api-access-9qkrh\") on node \"crc\" DevicePath \"\"" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.527491 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.527765 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.594792 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.652871 4979 generic.go:334] "Generic (PLEG): container finished" podID="575217c8-729c-4491-8d04-076ed4184afc" containerID="e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826" exitCode=0 Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.653464 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.653559 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" event={"ID":"575217c8-729c-4491-8d04-076ed4184afc","Type":"ContainerDied","Data":"e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826"} Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.653588 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d7b5456f5-847qc" event={"ID":"575217c8-729c-4491-8d04-076ed4184afc","Type":"ContainerDied","Data":"2376ee5772db2d3f62269eb7867e3d6f4f0ccdf6969787e28d4ab207e2f49303"} Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.653605 4979 scope.go:117] "RemoveContainer" containerID="e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.672500 4979 scope.go:117] "RemoveContainer" containerID="558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.689591 4979 scope.go:117] "RemoveContainer" containerID="e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826" Dec 04 13:02:20 crc kubenswrapper[4979]: E1204 13:02:20.691486 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826\": container with ID starting with e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826 not found: ID does not exist" containerID="e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.691527 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826"} err="failed to get container status \"e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826\": rpc error: code = NotFound desc = could not find container \"e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826\": container with ID starting with e5e157032e239fb40b67b90f342998db9249c424a9e66c364ad7696f2b387826 not found: ID does not exist" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.691553 4979 scope.go:117] "RemoveContainer" containerID="558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2" Dec 04 13:02:20 crc kubenswrapper[4979]: E1204 13:02:20.691894 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2\": container with ID starting with 558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2 not found: ID does not exist" containerID="558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.691931 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2"} err="failed to get container status \"558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2\": rpc error: code = NotFound desc = could not find container \"558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2\": container with ID starting with 558e4a082bb0462adabf4b3d8ef1ac587288685fe1076796116dcddd193e12c2 not found: ID does not exist" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.727241 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.889232 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc\") pod \"575217c8-729c-4491-8d04-076ed4184afc\" (UID: \"575217c8-729c-4491-8d04-076ed4184afc\") " Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.890079 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "575217c8-729c-4491-8d04-076ed4184afc" (UID: "575217c8-729c-4491-8d04-076ed4184afc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.981312 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-847qc"] Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.986898 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5d7b5456f5-847qc"] Dec 04 13:02:20 crc kubenswrapper[4979]: I1204 13:02:20.990508 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/575217c8-729c-4491-8d04-076ed4184afc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:02:21 crc kubenswrapper[4979]: I1204 13:02:21.465186 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:21 crc kubenswrapper[4979]: I1204 13:02:21.465258 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:22 crc kubenswrapper[4979]: I1204 13:02:22.216759 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="575217c8-729c-4491-8d04-076ed4184afc" path="/var/lib/kubelet/pods/575217c8-729c-4491-8d04-076ed4184afc/volumes" Dec 04 13:02:23 crc kubenswrapper[4979]: I1204 13:02:23.650836 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:23 crc kubenswrapper[4979]: I1204 13:02:23.721974 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 04 13:02:28 crc kubenswrapper[4979]: I1204 13:02:28.040698 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:02:28 crc kubenswrapper[4979]: I1204 13:02:28.041064 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:02:46 crc kubenswrapper[4979]: I1204 13:02:46.391566 4979 generic.go:334] "Generic (PLEG): container finished" podID="3e848001-e03d-4db6-84f9-24fecba801e4" containerID="ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4" exitCode=0 Dec 04 13:02:46 crc kubenswrapper[4979]: I1204 13:02:46.391629 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e848001-e03d-4db6-84f9-24fecba801e4","Type":"ContainerDied","Data":"ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4"} Dec 04 13:02:46 crc kubenswrapper[4979]: I1204 13:02:46.395068 4979 generic.go:334] "Generic (PLEG): container finished" podID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerID="db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6" exitCode=0 Dec 04 13:02:46 crc kubenswrapper[4979]: I1204 13:02:46.395110 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2","Type":"ContainerDied","Data":"db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6"} Dec 04 13:02:47 crc kubenswrapper[4979]: I1204 13:02:47.404404 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e848001-e03d-4db6-84f9-24fecba801e4","Type":"ContainerStarted","Data":"0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8"} Dec 04 13:02:47 crc kubenswrapper[4979]: I1204 13:02:47.404930 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 13:02:47 crc kubenswrapper[4979]: I1204 13:02:47.406879 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2","Type":"ContainerStarted","Data":"2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970"} Dec 04 13:02:47 crc kubenswrapper[4979]: I1204 13:02:47.407071 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:02:47 crc kubenswrapper[4979]: I1204 13:02:47.425612 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.42558997 podStartE2EDuration="39.42558997s" podCreationTimestamp="2025-12-04 13:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:47.425443145 +0000 UTC m=+4791.699738949" watchObservedRunningTime="2025-12-04 13:02:47.42558997 +0000 UTC m=+4791.699885774" Dec 04 13:02:47 crc kubenswrapper[4979]: I1204 13:02:47.454047 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.454026022 podStartE2EDuration="38.454026022s" podCreationTimestamp="2025-12-04 13:02:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:02:47.448243205 +0000 UTC m=+4791.722539019" watchObservedRunningTime="2025-12-04 13:02:47.454026022 +0000 UTC m=+4791.728321826" Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.040941 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.041709 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.041775 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.042657 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2cb5c9544ef934945129415b00a3831fb523499e2744b67a856f00423a013073"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.042760 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://2cb5c9544ef934945129415b00a3831fb523499e2744b67a856f00423a013073" gracePeriod=600 Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.496813 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="2cb5c9544ef934945129415b00a3831fb523499e2744b67a856f00423a013073" exitCode=0 Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.496863 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"2cb5c9544ef934945129415b00a3831fb523499e2744b67a856f00423a013073"} Dec 04 13:02:58 crc kubenswrapper[4979]: I1204 13:02:58.496894 4979 scope.go:117] "RemoveContainer" containerID="41e8aab8ca623b2ddce7878e74e69378b3a16ecb5666701933881b91e868a3d7" Dec 04 13:02:59 crc kubenswrapper[4979]: I1204 13:02:59.506886 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc"} Dec 04 13:03:00 crc kubenswrapper[4979]: I1204 13:03:00.094754 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 13:03:00 crc kubenswrapper[4979]: I1204 13:03:00.850475 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.107018 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-w5b8h"] Dec 04 13:03:04 crc kubenswrapper[4979]: E1204 13:03:04.108589 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575217c8-729c-4491-8d04-076ed4184afc" containerName="init" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.108610 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="575217c8-729c-4491-8d04-076ed4184afc" containerName="init" Dec 04 13:03:04 crc kubenswrapper[4979]: E1204 13:03:04.108638 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="575217c8-729c-4491-8d04-076ed4184afc" containerName="dnsmasq-dns" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.108646 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="575217c8-729c-4491-8d04-076ed4184afc" containerName="dnsmasq-dns" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.108875 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="575217c8-729c-4491-8d04-076ed4184afc" containerName="dnsmasq-dns" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.110248 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.128872 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-w5b8h"] Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.239062 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tx9x\" (UniqueName: \"kubernetes.io/projected/6c36cad8-31ac-477b-98f4-be4b4eeac17d-kube-api-access-9tx9x\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.240041 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.240068 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-config\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.341208 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.341257 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-config\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.341390 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tx9x\" (UniqueName: \"kubernetes.io/projected/6c36cad8-31ac-477b-98f4-be4b4eeac17d-kube-api-access-9tx9x\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.342146 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-dns-svc\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.342252 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-config\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.359151 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tx9x\" (UniqueName: \"kubernetes.io/projected/6c36cad8-31ac-477b-98f4-be4b4eeac17d-kube-api-access-9tx9x\") pod \"dnsmasq-dns-5b7946d7b9-w5b8h\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.437254 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.798555 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:03:04 crc kubenswrapper[4979]: W1204 13:03:04.857956 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c36cad8_31ac_477b_98f4_be4b4eeac17d.slice/crio-e11f6121e80629a53531bd50a35efc7593d7d5ec6810c6b4eea0ad6a872ec925 WatchSource:0}: Error finding container e11f6121e80629a53531bd50a35efc7593d7d5ec6810c6b4eea0ad6a872ec925: Status 404 returned error can't find the container with id e11f6121e80629a53531bd50a35efc7593d7d5ec6810c6b4eea0ad6a872ec925 Dec 04 13:03:04 crc kubenswrapper[4979]: I1204 13:03:04.860393 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-w5b8h"] Dec 04 13:03:05 crc kubenswrapper[4979]: I1204 13:03:05.451053 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:03:05 crc kubenswrapper[4979]: I1204 13:03:05.552149 4979 generic.go:334] "Generic (PLEG): container finished" podID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerID="62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4" exitCode=0 Dec 04 13:03:05 crc kubenswrapper[4979]: I1204 13:03:05.552201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" event={"ID":"6c36cad8-31ac-477b-98f4-be4b4eeac17d","Type":"ContainerDied","Data":"62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4"} Dec 04 13:03:05 crc kubenswrapper[4979]: I1204 13:03:05.552234 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" event={"ID":"6c36cad8-31ac-477b-98f4-be4b4eeac17d","Type":"ContainerStarted","Data":"e11f6121e80629a53531bd50a35efc7593d7d5ec6810c6b4eea0ad6a872ec925"} Dec 04 13:03:06 crc kubenswrapper[4979]: I1204 13:03:06.561802 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" event={"ID":"6c36cad8-31ac-477b-98f4-be4b4eeac17d","Type":"ContainerStarted","Data":"dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d"} Dec 04 13:03:06 crc kubenswrapper[4979]: I1204 13:03:06.562509 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:06 crc kubenswrapper[4979]: I1204 13:03:06.571530 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="rabbitmq" containerID="cri-o://0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8" gracePeriod=604799 Dec 04 13:03:06 crc kubenswrapper[4979]: I1204 13:03:06.585822 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" podStartSLOduration=2.585795468 podStartE2EDuration="2.585795468s" podCreationTimestamp="2025-12-04 13:03:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:03:06.580439932 +0000 UTC m=+4810.854735746" watchObservedRunningTime="2025-12-04 13:03:06.585795468 +0000 UTC m=+4810.860091272" Dec 04 13:03:07 crc kubenswrapper[4979]: I1204 13:03:07.141977 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="rabbitmq" containerID="cri-o://2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970" gracePeriod=604799 Dec 04 13:03:10 crc kubenswrapper[4979]: I1204 13:03:10.091212 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.241:5672: connect: connection refused" Dec 04 13:03:10 crc kubenswrapper[4979]: I1204 13:03:10.848648 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.244:5672: connect: connection refused" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.147197 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.281768 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-plugins\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.281947 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.281978 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-server-conf\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282028 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-plugins-conf\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282062 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e848001-e03d-4db6-84f9-24fecba801e4-erlang-cookie-secret\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282095 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-erlang-cookie\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282128 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6vwf\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-kube-api-access-p6vwf\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282153 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e848001-e03d-4db6-84f9-24fecba801e4-pod-info\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282153 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282211 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-confd\") pod \"3e848001-e03d-4db6-84f9-24fecba801e4\" (UID: \"3e848001-e03d-4db6-84f9-24fecba801e4\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282503 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282714 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.282911 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.287156 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e848001-e03d-4db6-84f9-24fecba801e4-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.288518 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-kube-api-access-p6vwf" (OuterVolumeSpecName: "kube-api-access-p6vwf") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "kube-api-access-p6vwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.290951 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3e848001-e03d-4db6-84f9-24fecba801e4-pod-info" (OuterVolumeSpecName: "pod-info") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.299134 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125" (OuterVolumeSpecName: "persistence") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "pvc-54998b18-2a27-4561-ac6a-fe6e22053125". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.303570 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-server-conf" (OuterVolumeSpecName: "server-conf") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.353656 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3e848001-e03d-4db6-84f9-24fecba801e4" (UID: "3e848001-e03d-4db6-84f9-24fecba801e4"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384049 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384092 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6vwf\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-kube-api-access-p6vwf\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384101 4979 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3e848001-e03d-4db6-84f9-24fecba801e4-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384110 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3e848001-e03d-4db6-84f9-24fecba801e4-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384156 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") on node \"crc\" " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384171 4979 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384181 4979 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3e848001-e03d-4db6-84f9-24fecba801e4-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.384190 4979 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3e848001-e03d-4db6-84f9-24fecba801e4-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.403707 4979 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.403952 4979 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-54998b18-2a27-4561-ac6a-fe6e22053125" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125") on node "crc" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.485398 4979 reconciler_common.go:293] "Volume detached for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.603207 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.616539 4979 generic.go:334] "Generic (PLEG): container finished" podID="3e848001-e03d-4db6-84f9-24fecba801e4" containerID="0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8" exitCode=0 Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.616621 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e848001-e03d-4db6-84f9-24fecba801e4","Type":"ContainerDied","Data":"0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8"} Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.616651 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3e848001-e03d-4db6-84f9-24fecba801e4","Type":"ContainerDied","Data":"cb885fc6a813660644d1a8caa6d508a64eef44e0d5f8c2b57397253da16d4278"} Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.616668 4979 scope.go:117] "RemoveContainer" containerID="0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.616792 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.622850 4979 generic.go:334] "Generic (PLEG): container finished" podID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerID="2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970" exitCode=0 Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.622909 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2","Type":"ContainerDied","Data":"2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970"} Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.622937 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2","Type":"ContainerDied","Data":"93cc9b08fdf7e1cd6d407a8be70d3473c0c88ee1161d9e358865e3272d7f0db3"} Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.623003 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.648746 4979 scope.go:117] "RemoveContainer" containerID="ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.674648 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687560 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-plugins-conf\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687635 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-confd\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687657 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-plugins\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687679 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-erlang-cookie\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687706 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-pod-info\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687730 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-erlang-cookie-secret\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687769 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-server-conf\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687800 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9898\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-kube-api-access-s9898\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.687927 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\" (UID: \"673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2\") " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.688904 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.689165 4979 scope.go:117] "RemoveContainer" containerID="0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.690078 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8\": container with ID starting with 0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8 not found: ID does not exist" containerID="0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.690126 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8"} err="failed to get container status \"0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8\": rpc error: code = NotFound desc = could not find container \"0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8\": container with ID starting with 0b9374a8f78420f2d975bdbd9bad161493363616b196be2d441febe68c39e1b8 not found: ID does not exist" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.690155 4979 scope.go:117] "RemoveContainer" containerID="ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.690170 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.690519 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.690828 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4\": container with ID starting with ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4 not found: ID does not exist" containerID="ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.690876 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4"} err="failed to get container status \"ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4\": rpc error: code = NotFound desc = could not find container \"ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4\": container with ID starting with ae41da230e92bbea32a3f1af27a92066ee1a57df5a5de1551fa94bafee1096b4 not found: ID does not exist" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.690891 4979 scope.go:117] "RemoveContainer" containerID="2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.694109 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.694132 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-pod-info" (OuterVolumeSpecName: "pod-info") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.694190 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.697071 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-kube-api-access-s9898" (OuterVolumeSpecName: "kube-api-access-s9898") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "kube-api-access-s9898". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.715084 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-server-conf" (OuterVolumeSpecName: "server-conf") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.720136 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc" (OuterVolumeSpecName: "persistence") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.720663 4979 scope.go:117] "RemoveContainer" containerID="db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.725835 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.726182 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="rabbitmq" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.726201 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="rabbitmq" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.726216 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="setup-container" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.726222 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="setup-container" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.726242 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="rabbitmq" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.726249 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="rabbitmq" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.726270 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="setup-container" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.726275 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="setup-container" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.726416 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" containerName="rabbitmq" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.726441 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" containerName="rabbitmq" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.727229 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.729417 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-wbpgc" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.729721 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.729874 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.730021 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.731835 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.733252 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.744242 4979 scope.go:117] "RemoveContainer" containerID="2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.745331 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970\": container with ID starting with 2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970 not found: ID does not exist" containerID="2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.745374 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970"} err="failed to get container status \"2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970\": rpc error: code = NotFound desc = could not find container \"2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970\": container with ID starting with 2ea3be0b61651ddf0f2854dbfcd2f30f6682aa66e5622554ff4dd52967026970 not found: ID does not exist" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.745410 4979 scope.go:117] "RemoveContainer" containerID="db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6" Dec 04 13:03:13 crc kubenswrapper[4979]: E1204 13:03:13.748825 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6\": container with ID starting with db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6 not found: ID does not exist" containerID="db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.748860 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6"} err="failed to get container status \"db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6\": rpc error: code = NotFound desc = could not find container \"db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6\": container with ID starting with db21d535ad8524c033c5ae7285bd159bc2680cfefc5ab174f15fe8f83651e0e6 not found: ID does not exist" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.786453 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" (UID: "673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789700 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2149bf59-192d-42c6-98a8-989643a5094b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789744 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789766 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2149bf59-192d-42c6-98a8-989643a5094b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789798 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789825 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789842 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2149bf59-192d-42c6-98a8-989643a5094b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789858 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789881 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2149bf59-192d-42c6-98a8-989643a5094b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789914 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8cll\" (UniqueName: \"kubernetes.io/projected/2149bf59-192d-42c6-98a8-989643a5094b-kube-api-access-v8cll\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789969 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789982 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.789992 4979 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.790002 4979 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-pod-info\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.790010 4979 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.790017 4979 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-server-conf\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.790025 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9898\" (UniqueName: \"kubernetes.io/projected/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-kube-api-access-s9898\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.790046 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") on node \"crc\" " Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.790056 4979 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.809010 4979 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.809167 4979 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc") on node "crc" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890769 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8cll\" (UniqueName: \"kubernetes.io/projected/2149bf59-192d-42c6-98a8-989643a5094b-kube-api-access-v8cll\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890846 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2149bf59-192d-42c6-98a8-989643a5094b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890880 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890901 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2149bf59-192d-42c6-98a8-989643a5094b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890930 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890956 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.890969 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2149bf59-192d-42c6-98a8-989643a5094b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.891015 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.891033 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2149bf59-192d-42c6-98a8-989643a5094b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.891084 4979 reconciler_common.go:293] "Volume detached for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.891815 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2149bf59-192d-42c6-98a8-989643a5094b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.892273 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.893380 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.893557 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2149bf59-192d-42c6-98a8-989643a5094b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.895288 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2149bf59-192d-42c6-98a8-989643a5094b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.897370 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2149bf59-192d-42c6-98a8-989643a5094b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.898182 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2149bf59-192d-42c6-98a8-989643a5094b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.901599 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.901711 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2be5edbcc6da3df1b51063102b8f2463830666ce99ffb0e08566e12b8abd0f29/globalmount\"" pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.908843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8cll\" (UniqueName: \"kubernetes.io/projected/2149bf59-192d-42c6-98a8-989643a5094b-kube-api-access-v8cll\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.944703 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-54998b18-2a27-4561-ac6a-fe6e22053125\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-54998b18-2a27-4561-ac6a-fe6e22053125\") pod \"rabbitmq-server-0\" (UID: \"2149bf59-192d-42c6-98a8-989643a5094b\") " pod="openstack/rabbitmq-server-0" Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.977449 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:03:13 crc kubenswrapper[4979]: I1204 13:03:13.981759 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.003913 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.011643 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.017852 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.017934 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.017972 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.017941 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-2gx5p" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.018060 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.020336 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.049578 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.098423 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.098583 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.098712 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.098816 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.099003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.099119 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh25v\" (UniqueName: \"kubernetes.io/projected/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-kube-api-access-xh25v\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.099166 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.099188 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.099235 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201780 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh25v\" (UniqueName: \"kubernetes.io/projected/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-kube-api-access-xh25v\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201823 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201854 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201873 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201941 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.201965 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.202166 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.202815 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.203003 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.203553 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.203891 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.204383 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.212417 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.212421 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.212579 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c04b0c08aa6505c08e21eaa4ffb138669ded1689a8c37464271f4824abcccd00/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.213009 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.216821 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.217836 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e848001-e03d-4db6-84f9-24fecba801e4" path="/var/lib/kubelet/pods/3e848001-e03d-4db6-84f9-24fecba801e4/volumes" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.219178 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2" path="/var/lib/kubelet/pods/673d7a64-3fbe-4fd8-9dae-d7b2a3edd0c2/volumes" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.224149 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh25v\" (UniqueName: \"kubernetes.io/projected/61f5d1b2-dc4a-49f5-8c2c-ce7193287348-kube-api-access-xh25v\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.243994 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5f28db56-b54c-4df5-86d5-1f441d48cdfc\") pod \"rabbitmq-cell1-server-0\" (UID: \"61f5d1b2-dc4a-49f5-8c2c-ce7193287348\") " pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.327978 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.439496 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.502741 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.518091 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-7xlrw"] Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.518321 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerName="dnsmasq-dns" containerID="cri-o://2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238" gracePeriod=10 Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.639118 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2149bf59-192d-42c6-98a8-989643a5094b","Type":"ContainerStarted","Data":"12ec18813b53bd718173ec48f6ccb2b66545a6c8fa2e93f2a4bf3508c9d70d6f"} Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.805868 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 04 13:03:14 crc kubenswrapper[4979]: I1204 13:03:14.938919 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.017599 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppgs5\" (UniqueName: \"kubernetes.io/projected/8bf461b9-e97a-46ff-95c3-6f32f568e87f-kube-api-access-ppgs5\") pod \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.018095 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-dns-svc\") pod \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.018139 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-config\") pod \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\" (UID: \"8bf461b9-e97a-46ff-95c3-6f32f568e87f\") " Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.029413 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf461b9-e97a-46ff-95c3-6f32f568e87f-kube-api-access-ppgs5" (OuterVolumeSpecName: "kube-api-access-ppgs5") pod "8bf461b9-e97a-46ff-95c3-6f32f568e87f" (UID: "8bf461b9-e97a-46ff-95c3-6f32f568e87f"). InnerVolumeSpecName "kube-api-access-ppgs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.052927 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8bf461b9-e97a-46ff-95c3-6f32f568e87f" (UID: "8bf461b9-e97a-46ff-95c3-6f32f568e87f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.053632 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-config" (OuterVolumeSpecName: "config") pod "8bf461b9-e97a-46ff-95c3-6f32f568e87f" (UID: "8bf461b9-e97a-46ff-95c3-6f32f568e87f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.120563 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppgs5\" (UniqueName: \"kubernetes.io/projected/8bf461b9-e97a-46ff-95c3-6f32f568e87f-kube-api-access-ppgs5\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.120608 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.120623 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8bf461b9-e97a-46ff-95c3-6f32f568e87f-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.652071 4979 generic.go:334] "Generic (PLEG): container finished" podID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerID="2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238" exitCode=0 Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.652131 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.652121 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" event={"ID":"8bf461b9-e97a-46ff-95c3-6f32f568e87f","Type":"ContainerDied","Data":"2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238"} Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.652265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-98ddfc8f-7xlrw" event={"ID":"8bf461b9-e97a-46ff-95c3-6f32f568e87f","Type":"ContainerDied","Data":"bef695a232dd93976ddf75974b6d2a9a461fb07cd80c40dcf7eb626dbb25762f"} Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.652315 4979 scope.go:117] "RemoveContainer" containerID="2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.653787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61f5d1b2-dc4a-49f5-8c2c-ce7193287348","Type":"ContainerStarted","Data":"20fae6805e84dfbffc912390c996143ec0937c719afd3ee992b6a95c5d015473"} Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.672658 4979 scope.go:117] "RemoveContainer" containerID="2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.693835 4979 scope.go:117] "RemoveContainer" containerID="2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238" Dec 04 13:03:15 crc kubenswrapper[4979]: E1204 13:03:15.694323 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238\": container with ID starting with 2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238 not found: ID does not exist" containerID="2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.694358 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238"} err="failed to get container status \"2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238\": rpc error: code = NotFound desc = could not find container \"2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238\": container with ID starting with 2724a1fbeb009411f506324552d5e8ba2e9d3cac04b07be9c947d03b97293238 not found: ID does not exist" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.694382 4979 scope.go:117] "RemoveContainer" containerID="2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c" Dec 04 13:03:15 crc kubenswrapper[4979]: E1204 13:03:15.694668 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c\": container with ID starting with 2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c not found: ID does not exist" containerID="2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.694701 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c"} err="failed to get container status \"2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c\": rpc error: code = NotFound desc = could not find container \"2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c\": container with ID starting with 2e6a1944fda5d4ef48541689cc592374af79bc867b66ac71bdba7b9a29c7908c not found: ID does not exist" Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.698479 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-7xlrw"] Dec 04 13:03:15 crc kubenswrapper[4979]: I1204 13:03:15.705930 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-98ddfc8f-7xlrw"] Dec 04 13:03:16 crc kubenswrapper[4979]: I1204 13:03:16.210643 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" path="/var/lib/kubelet/pods/8bf461b9-e97a-46ff-95c3-6f32f568e87f/volumes" Dec 04 13:03:16 crc kubenswrapper[4979]: I1204 13:03:16.663222 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2149bf59-192d-42c6-98a8-989643a5094b","Type":"ContainerStarted","Data":"cee689875c2b41441ccee3048d1790040830f408e1ba72a1fd990907cf398bad"} Dec 04 13:03:16 crc kubenswrapper[4979]: I1204 13:03:16.664821 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61f5d1b2-dc4a-49f5-8c2c-ce7193287348","Type":"ContainerStarted","Data":"4b703c6c9826435ba8079ec41f2c096ba9516419bef3b113d33e8fea7d4aa5b0"} Dec 04 13:03:48 crc kubenswrapper[4979]: I1204 13:03:48.909005 4979 generic.go:334] "Generic (PLEG): container finished" podID="2149bf59-192d-42c6-98a8-989643a5094b" containerID="cee689875c2b41441ccee3048d1790040830f408e1ba72a1fd990907cf398bad" exitCode=0 Dec 04 13:03:48 crc kubenswrapper[4979]: I1204 13:03:48.909140 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2149bf59-192d-42c6-98a8-989643a5094b","Type":"ContainerDied","Data":"cee689875c2b41441ccee3048d1790040830f408e1ba72a1fd990907cf398bad"} Dec 04 13:03:48 crc kubenswrapper[4979]: I1204 13:03:48.912656 4979 generic.go:334] "Generic (PLEG): container finished" podID="61f5d1b2-dc4a-49f5-8c2c-ce7193287348" containerID="4b703c6c9826435ba8079ec41f2c096ba9516419bef3b113d33e8fea7d4aa5b0" exitCode=0 Dec 04 13:03:48 crc kubenswrapper[4979]: I1204 13:03:48.912736 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61f5d1b2-dc4a-49f5-8c2c-ce7193287348","Type":"ContainerDied","Data":"4b703c6c9826435ba8079ec41f2c096ba9516419bef3b113d33e8fea7d4aa5b0"} Dec 04 13:03:49 crc kubenswrapper[4979]: I1204 13:03:49.923892 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"61f5d1b2-dc4a-49f5-8c2c-ce7193287348","Type":"ContainerStarted","Data":"d27784b9d6b0e7aefee7820db38d068ceebf431d43ac8ea76e38d6d1d0722f3c"} Dec 04 13:03:49 crc kubenswrapper[4979]: I1204 13:03:49.924573 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:03:49 crc kubenswrapper[4979]: I1204 13:03:49.926634 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2149bf59-192d-42c6-98a8-989643a5094b","Type":"ContainerStarted","Data":"70bf5491515520907a5c31cca61a42f68affce7db36aa41ff18caf3cff8fcc33"} Dec 04 13:03:49 crc kubenswrapper[4979]: I1204 13:03:49.926866 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 04 13:03:49 crc kubenswrapper[4979]: I1204 13:03:49.955425 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.955404629 podStartE2EDuration="36.955404629s" podCreationTimestamp="2025-12-04 13:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:03:49.94769285 +0000 UTC m=+4854.221988684" watchObservedRunningTime="2025-12-04 13:03:49.955404629 +0000 UTC m=+4854.229700443" Dec 04 13:03:49 crc kubenswrapper[4979]: I1204 13:03:49.976631 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.976611425 podStartE2EDuration="36.976611425s" podCreationTimestamp="2025-12-04 13:03:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:03:49.972579075 +0000 UTC m=+4854.246874889" watchObservedRunningTime="2025-12-04 13:03:49.976611425 +0000 UTC m=+4854.250907229" Dec 04 13:04:04 crc kubenswrapper[4979]: I1204 13:04:04.052544 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 04 13:04:04 crc kubenswrapper[4979]: I1204 13:04:04.330946 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.855557 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 13:04:10 crc kubenswrapper[4979]: E1204 13:04:10.856377 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerName="dnsmasq-dns" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.856389 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerName="dnsmasq-dns" Dec 04 13:04:10 crc kubenswrapper[4979]: E1204 13:04:10.856410 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerName="init" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.856416 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerName="init" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.856567 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf461b9-e97a-46ff-95c3-6f32f568e87f" containerName="dnsmasq-dns" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.857096 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.864884 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.867231 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-wx59x" Dec 04 13:04:10 crc kubenswrapper[4979]: I1204 13:04:10.976107 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc2xh\" (UniqueName: \"kubernetes.io/projected/c7e0c8b1-bd69-49ca-933d-4c8776b24605-kube-api-access-cc2xh\") pod \"mariadb-client-1-default\" (UID: \"c7e0c8b1-bd69-49ca-933d-4c8776b24605\") " pod="openstack/mariadb-client-1-default" Dec 04 13:04:11 crc kubenswrapper[4979]: I1204 13:04:11.078350 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc2xh\" (UniqueName: \"kubernetes.io/projected/c7e0c8b1-bd69-49ca-933d-4c8776b24605-kube-api-access-cc2xh\") pod \"mariadb-client-1-default\" (UID: \"c7e0c8b1-bd69-49ca-933d-4c8776b24605\") " pod="openstack/mariadb-client-1-default" Dec 04 13:04:11 crc kubenswrapper[4979]: I1204 13:04:11.096944 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc2xh\" (UniqueName: \"kubernetes.io/projected/c7e0c8b1-bd69-49ca-933d-4c8776b24605-kube-api-access-cc2xh\") pod \"mariadb-client-1-default\" (UID: \"c7e0c8b1-bd69-49ca-933d-4c8776b24605\") " pod="openstack/mariadb-client-1-default" Dec 04 13:04:11 crc kubenswrapper[4979]: I1204 13:04:11.174237 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 13:04:11 crc kubenswrapper[4979]: I1204 13:04:11.683682 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 13:04:11 crc kubenswrapper[4979]: W1204 13:04:11.688338 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7e0c8b1_bd69_49ca_933d_4c8776b24605.slice/crio-7e857118e534fdbc8f79afb7f9ec1da051a9bcd317334be86e6ccaebd9e39859 WatchSource:0}: Error finding container 7e857118e534fdbc8f79afb7f9ec1da051a9bcd317334be86e6ccaebd9e39859: Status 404 returned error can't find the container with id 7e857118e534fdbc8f79afb7f9ec1da051a9bcd317334be86e6ccaebd9e39859 Dec 04 13:04:12 crc kubenswrapper[4979]: I1204 13:04:12.094178 4979 generic.go:334] "Generic (PLEG): container finished" podID="c7e0c8b1-bd69-49ca-933d-4c8776b24605" containerID="8dedfa414fac3e373b31ea9d5fd3e57c8404a857c25566864a85c31dd85d7e91" exitCode=0 Dec 04 13:04:12 crc kubenswrapper[4979]: I1204 13:04:12.094288 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"c7e0c8b1-bd69-49ca-933d-4c8776b24605","Type":"ContainerDied","Data":"8dedfa414fac3e373b31ea9d5fd3e57c8404a857c25566864a85c31dd85d7e91"} Dec 04 13:04:12 crc kubenswrapper[4979]: I1204 13:04:12.094497 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1-default" event={"ID":"c7e0c8b1-bd69-49ca-933d-4c8776b24605","Type":"ContainerStarted","Data":"7e857118e534fdbc8f79afb7f9ec1da051a9bcd317334be86e6ccaebd9e39859"} Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.404672 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.429879 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1-default_c7e0c8b1-bd69-49ca-933d-4c8776b24605/mariadb-client-1-default/0.log" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.457083 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.462107 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1-default"] Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.517493 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc2xh\" (UniqueName: \"kubernetes.io/projected/c7e0c8b1-bd69-49ca-933d-4c8776b24605-kube-api-access-cc2xh\") pod \"c7e0c8b1-bd69-49ca-933d-4c8776b24605\" (UID: \"c7e0c8b1-bd69-49ca-933d-4c8776b24605\") " Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.523441 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7e0c8b1-bd69-49ca-933d-4c8776b24605-kube-api-access-cc2xh" (OuterVolumeSpecName: "kube-api-access-cc2xh") pod "c7e0c8b1-bd69-49ca-933d-4c8776b24605" (UID: "c7e0c8b1-bd69-49ca-933d-4c8776b24605"). InnerVolumeSpecName "kube-api-access-cc2xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.619385 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc2xh\" (UniqueName: \"kubernetes.io/projected/c7e0c8b1-bd69-49ca-933d-4c8776b24605-kube-api-access-cc2xh\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.880489 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 13:04:13 crc kubenswrapper[4979]: E1204 13:04:13.880793 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7e0c8b1-bd69-49ca-933d-4c8776b24605" containerName="mariadb-client-1-default" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.880804 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7e0c8b1-bd69-49ca-933d-4c8776b24605" containerName="mariadb-client-1-default" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.880987 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7e0c8b1-bd69-49ca-933d-4c8776b24605" containerName="mariadb-client-1-default" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.881730 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.890017 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 13:04:13 crc kubenswrapper[4979]: I1204 13:04:13.924129 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkp7g\" (UniqueName: \"kubernetes.io/projected/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3-kube-api-access-qkp7g\") pod \"mariadb-client-2-default\" (UID: \"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3\") " pod="openstack/mariadb-client-2-default" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.026201 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkp7g\" (UniqueName: \"kubernetes.io/projected/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3-kube-api-access-qkp7g\") pod \"mariadb-client-2-default\" (UID: \"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3\") " pod="openstack/mariadb-client-2-default" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.048245 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkp7g\" (UniqueName: \"kubernetes.io/projected/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3-kube-api-access-qkp7g\") pod \"mariadb-client-2-default\" (UID: \"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3\") " pod="openstack/mariadb-client-2-default" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.111551 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e857118e534fdbc8f79afb7f9ec1da051a9bcd317334be86e6ccaebd9e39859" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.111659 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1-default" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.202223 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.209438 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7e0c8b1-bd69-49ca-933d-4c8776b24605" path="/var/lib/kubelet/pods/c7e0c8b1-bd69-49ca-933d-4c8776b24605/volumes" Dec 04 13:04:14 crc kubenswrapper[4979]: I1204 13:04:14.480161 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 13:04:14 crc kubenswrapper[4979]: W1204 13:04:14.491683 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fb0fbd5_016a_4346_8c1d_cc59a726a2a3.slice/crio-3a0b6719d34228d9389c73452bd7fc44f765a52c98841a94a005c1c08b21816e WatchSource:0}: Error finding container 3a0b6719d34228d9389c73452bd7fc44f765a52c98841a94a005c1c08b21816e: Status 404 returned error can't find the container with id 3a0b6719d34228d9389c73452bd7fc44f765a52c98841a94a005c1c08b21816e Dec 04 13:04:15 crc kubenswrapper[4979]: I1204 13:04:15.119065 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3","Type":"ContainerStarted","Data":"f0beb506bad144e6cc3bac4ccbcae612173f3b965f8d3f345a419c4bac2779fe"} Dec 04 13:04:15 crc kubenswrapper[4979]: I1204 13:04:15.119371 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3","Type":"ContainerStarted","Data":"3a0b6719d34228d9389c73452bd7fc44f765a52c98841a94a005c1c08b21816e"} Dec 04 13:04:15 crc kubenswrapper[4979]: I1204 13:04:15.136390 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client-2-default" podStartSLOduration=2.136369259 podStartE2EDuration="2.136369259s" podCreationTimestamp="2025-12-04 13:04:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:04:15.131011354 +0000 UTC m=+4879.405307158" watchObservedRunningTime="2025-12-04 13:04:15.136369259 +0000 UTC m=+4879.410665063" Dec 04 13:04:16 crc kubenswrapper[4979]: I1204 13:04:16.128332 4979 generic.go:334] "Generic (PLEG): container finished" podID="1fb0fbd5-016a-4346-8c1d-cc59a726a2a3" containerID="f0beb506bad144e6cc3bac4ccbcae612173f3b965f8d3f345a419c4bac2779fe" exitCode=1 Dec 04 13:04:16 crc kubenswrapper[4979]: I1204 13:04:16.128452 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2-default" event={"ID":"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3","Type":"ContainerDied","Data":"f0beb506bad144e6cc3bac4ccbcae612173f3b965f8d3f345a419c4bac2779fe"} Dec 04 13:04:17 crc kubenswrapper[4979]: I1204 13:04:17.511725 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 13:04:17 crc kubenswrapper[4979]: I1204 13:04:17.554230 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 13:04:17 crc kubenswrapper[4979]: I1204 13:04:17.560289 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2-default"] Dec 04 13:04:17 crc kubenswrapper[4979]: I1204 13:04:17.575884 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkp7g\" (UniqueName: \"kubernetes.io/projected/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3-kube-api-access-qkp7g\") pod \"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3\" (UID: \"1fb0fbd5-016a-4346-8c1d-cc59a726a2a3\") " Dec 04 13:04:17 crc kubenswrapper[4979]: I1204 13:04:17.580711 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3-kube-api-access-qkp7g" (OuterVolumeSpecName: "kube-api-access-qkp7g") pod "1fb0fbd5-016a-4346-8c1d-cc59a726a2a3" (UID: "1fb0fbd5-016a-4346-8c1d-cc59a726a2a3"). InnerVolumeSpecName "kube-api-access-qkp7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:17 crc kubenswrapper[4979]: I1204 13:04:17.677824 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkp7g\" (UniqueName: \"kubernetes.io/projected/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3-kube-api-access-qkp7g\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.102120 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-1"] Dec 04 13:04:18 crc kubenswrapper[4979]: E1204 13:04:18.102911 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fb0fbd5-016a-4346-8c1d-cc59a726a2a3" containerName="mariadb-client-2-default" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.103042 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fb0fbd5-016a-4346-8c1d-cc59a726a2a3" containerName="mariadb-client-2-default" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.103265 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fb0fbd5-016a-4346-8c1d-cc59a726a2a3" containerName="mariadb-client-2-default" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.103972 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.112687 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.146431 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a0b6719d34228d9389c73452bd7fc44f765a52c98841a94a005c1c08b21816e" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.146500 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2-default" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.185192 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5lst\" (UniqueName: \"kubernetes.io/projected/6ebb4a49-2c87-434a-bc3d-6c35873a72ef-kube-api-access-v5lst\") pod \"mariadb-client-1\" (UID: \"6ebb4a49-2c87-434a-bc3d-6c35873a72ef\") " pod="openstack/mariadb-client-1" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.210684 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fb0fbd5-016a-4346-8c1d-cc59a726a2a3" path="/var/lib/kubelet/pods/1fb0fbd5-016a-4346-8c1d-cc59a726a2a3/volumes" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.286746 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5lst\" (UniqueName: \"kubernetes.io/projected/6ebb4a49-2c87-434a-bc3d-6c35873a72ef-kube-api-access-v5lst\") pod \"mariadb-client-1\" (UID: \"6ebb4a49-2c87-434a-bc3d-6c35873a72ef\") " pod="openstack/mariadb-client-1" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.308108 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5lst\" (UniqueName: \"kubernetes.io/projected/6ebb4a49-2c87-434a-bc3d-6c35873a72ef-kube-api-access-v5lst\") pod \"mariadb-client-1\" (UID: \"6ebb4a49-2c87-434a-bc3d-6c35873a72ef\") " pod="openstack/mariadb-client-1" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.427541 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 13:04:18 crc kubenswrapper[4979]: I1204 13:04:18.959845 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 13:04:18 crc kubenswrapper[4979]: W1204 13:04:18.965490 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ebb4a49_2c87_434a_bc3d_6c35873a72ef.slice/crio-7af8463cf6c1d8eb5b0c454c72dac119f4bcf314d8f8e11940d5903f4fe7edff WatchSource:0}: Error finding container 7af8463cf6c1d8eb5b0c454c72dac119f4bcf314d8f8e11940d5903f4fe7edff: Status 404 returned error can't find the container with id 7af8463cf6c1d8eb5b0c454c72dac119f4bcf314d8f8e11940d5903f4fe7edff Dec 04 13:04:19 crc kubenswrapper[4979]: I1204 13:04:19.155114 4979 generic.go:334] "Generic (PLEG): container finished" podID="6ebb4a49-2c87-434a-bc3d-6c35873a72ef" containerID="d3671dfc11e3788e515d961c21fbb488d7416ca55d5634f6eee5e24055d5ba67" exitCode=0 Dec 04 13:04:19 crc kubenswrapper[4979]: I1204 13:04:19.155159 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"6ebb4a49-2c87-434a-bc3d-6c35873a72ef","Type":"ContainerDied","Data":"d3671dfc11e3788e515d961c21fbb488d7416ca55d5634f6eee5e24055d5ba67"} Dec 04 13:04:19 crc kubenswrapper[4979]: I1204 13:04:19.155182 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-1" event={"ID":"6ebb4a49-2c87-434a-bc3d-6c35873a72ef","Type":"ContainerStarted","Data":"7af8463cf6c1d8eb5b0c454c72dac119f4bcf314d8f8e11940d5903f4fe7edff"} Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.523627 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.547663 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-1_6ebb4a49-2c87-434a-bc3d-6c35873a72ef/mariadb-client-1/0.log" Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.574741 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.582125 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-1"] Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.623147 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5lst\" (UniqueName: \"kubernetes.io/projected/6ebb4a49-2c87-434a-bc3d-6c35873a72ef-kube-api-access-v5lst\") pod \"6ebb4a49-2c87-434a-bc3d-6c35873a72ef\" (UID: \"6ebb4a49-2c87-434a-bc3d-6c35873a72ef\") " Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.628798 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ebb4a49-2c87-434a-bc3d-6c35873a72ef-kube-api-access-v5lst" (OuterVolumeSpecName: "kube-api-access-v5lst") pod "6ebb4a49-2c87-434a-bc3d-6c35873a72ef" (UID: "6ebb4a49-2c87-434a-bc3d-6c35873a72ef"). InnerVolumeSpecName "kube-api-access-v5lst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:20 crc kubenswrapper[4979]: I1204 13:04:20.725014 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5lst\" (UniqueName: \"kubernetes.io/projected/6ebb4a49-2c87-434a-bc3d-6c35873a72ef-kube-api-access-v5lst\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.026533 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 13:04:21 crc kubenswrapper[4979]: E1204 13:04:21.027360 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ebb4a49-2c87-434a-bc3d-6c35873a72ef" containerName="mariadb-client-1" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.027376 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ebb4a49-2c87-434a-bc3d-6c35873a72ef" containerName="mariadb-client-1" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.027523 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ebb4a49-2c87-434a-bc3d-6c35873a72ef" containerName="mariadb-client-1" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.028099 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.037790 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.132012 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt9br\" (UniqueName: \"kubernetes.io/projected/60e86772-aed2-4c8a-8a61-86f0e0886efb-kube-api-access-kt9br\") pod \"mariadb-client-4-default\" (UID: \"60e86772-aed2-4c8a-8a61-86f0e0886efb\") " pod="openstack/mariadb-client-4-default" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.171545 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7af8463cf6c1d8eb5b0c454c72dac119f4bcf314d8f8e11940d5903f4fe7edff" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.171624 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-1" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.233536 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt9br\" (UniqueName: \"kubernetes.io/projected/60e86772-aed2-4c8a-8a61-86f0e0886efb-kube-api-access-kt9br\") pod \"mariadb-client-4-default\" (UID: \"60e86772-aed2-4c8a-8a61-86f0e0886efb\") " pod="openstack/mariadb-client-4-default" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.251274 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt9br\" (UniqueName: \"kubernetes.io/projected/60e86772-aed2-4c8a-8a61-86f0e0886efb-kube-api-access-kt9br\") pod \"mariadb-client-4-default\" (UID: \"60e86772-aed2-4c8a-8a61-86f0e0886efb\") " pod="openstack/mariadb-client-4-default" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.359982 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 13:04:21 crc kubenswrapper[4979]: I1204 13:04:21.869837 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 13:04:21 crc kubenswrapper[4979]: W1204 13:04:21.875644 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60e86772_aed2_4c8a_8a61_86f0e0886efb.slice/crio-74e53156bf552ffb15b8ca2a29ea4831b98c4232397c614057fe2bdc974a9686 WatchSource:0}: Error finding container 74e53156bf552ffb15b8ca2a29ea4831b98c4232397c614057fe2bdc974a9686: Status 404 returned error can't find the container with id 74e53156bf552ffb15b8ca2a29ea4831b98c4232397c614057fe2bdc974a9686 Dec 04 13:04:22 crc kubenswrapper[4979]: I1204 13:04:22.183257 4979 generic.go:334] "Generic (PLEG): container finished" podID="60e86772-aed2-4c8a-8a61-86f0e0886efb" containerID="75a41cb2318ca07ac1de7c68dabd2ec60020f005f0bb882f540537a7c1446efe" exitCode=0 Dec 04 13:04:22 crc kubenswrapper[4979]: I1204 13:04:22.183349 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"60e86772-aed2-4c8a-8a61-86f0e0886efb","Type":"ContainerDied","Data":"75a41cb2318ca07ac1de7c68dabd2ec60020f005f0bb882f540537a7c1446efe"} Dec 04 13:04:22 crc kubenswrapper[4979]: I1204 13:04:22.183666 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-4-default" event={"ID":"60e86772-aed2-4c8a-8a61-86f0e0886efb","Type":"ContainerStarted","Data":"74e53156bf552ffb15b8ca2a29ea4831b98c4232397c614057fe2bdc974a9686"} Dec 04 13:04:22 crc kubenswrapper[4979]: I1204 13:04:22.211146 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ebb4a49-2c87-434a-bc3d-6c35873a72ef" path="/var/lib/kubelet/pods/6ebb4a49-2c87-434a-bc3d-6c35873a72ef/volumes" Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.540019 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.560514 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-4-default_60e86772-aed2-4c8a-8a61-86f0e0886efb/mariadb-client-4-default/0.log" Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.585585 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.593354 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-4-default"] Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.674637 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kt9br\" (UniqueName: \"kubernetes.io/projected/60e86772-aed2-4c8a-8a61-86f0e0886efb-kube-api-access-kt9br\") pod \"60e86772-aed2-4c8a-8a61-86f0e0886efb\" (UID: \"60e86772-aed2-4c8a-8a61-86f0e0886efb\") " Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.680335 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60e86772-aed2-4c8a-8a61-86f0e0886efb-kube-api-access-kt9br" (OuterVolumeSpecName: "kube-api-access-kt9br") pod "60e86772-aed2-4c8a-8a61-86f0e0886efb" (UID: "60e86772-aed2-4c8a-8a61-86f0e0886efb"). InnerVolumeSpecName "kube-api-access-kt9br". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:23 crc kubenswrapper[4979]: I1204 13:04:23.776853 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kt9br\" (UniqueName: \"kubernetes.io/projected/60e86772-aed2-4c8a-8a61-86f0e0886efb-kube-api-access-kt9br\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:24 crc kubenswrapper[4979]: I1204 13:04:24.203643 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-4-default" Dec 04 13:04:24 crc kubenswrapper[4979]: I1204 13:04:24.215940 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60e86772-aed2-4c8a-8a61-86f0e0886efb" path="/var/lib/kubelet/pods/60e86772-aed2-4c8a-8a61-86f0e0886efb/volumes" Dec 04 13:04:24 crc kubenswrapper[4979]: I1204 13:04:24.216569 4979 scope.go:117] "RemoveContainer" containerID="75a41cb2318ca07ac1de7c68dabd2ec60020f005f0bb882f540537a7c1446efe" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.089522 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 13:04:27 crc kubenswrapper[4979]: E1204 13:04:27.090417 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60e86772-aed2-4c8a-8a61-86f0e0886efb" containerName="mariadb-client-4-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.090442 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="60e86772-aed2-4c8a-8a61-86f0e0886efb" containerName="mariadb-client-4-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.090686 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="60e86772-aed2-4c8a-8a61-86f0e0886efb" containerName="mariadb-client-4-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.091472 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.093539 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-wx59x" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.097603 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.230352 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb7sj\" (UniqueName: \"kubernetes.io/projected/042bf5d7-9b0d-4ee7-84eb-c6209a607717-kube-api-access-jb7sj\") pod \"mariadb-client-5-default\" (UID: \"042bf5d7-9b0d-4ee7-84eb-c6209a607717\") " pod="openstack/mariadb-client-5-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.331665 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb7sj\" (UniqueName: \"kubernetes.io/projected/042bf5d7-9b0d-4ee7-84eb-c6209a607717-kube-api-access-jb7sj\") pod \"mariadb-client-5-default\" (UID: \"042bf5d7-9b0d-4ee7-84eb-c6209a607717\") " pod="openstack/mariadb-client-5-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.354053 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb7sj\" (UniqueName: \"kubernetes.io/projected/042bf5d7-9b0d-4ee7-84eb-c6209a607717-kube-api-access-jb7sj\") pod \"mariadb-client-5-default\" (UID: \"042bf5d7-9b0d-4ee7-84eb-c6209a607717\") " pod="openstack/mariadb-client-5-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.410195 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 13:04:27 crc kubenswrapper[4979]: I1204 13:04:27.918712 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 13:04:28 crc kubenswrapper[4979]: I1204 13:04:28.236485 4979 generic.go:334] "Generic (PLEG): container finished" podID="042bf5d7-9b0d-4ee7-84eb-c6209a607717" containerID="a5a70153cf74ece77af161689936ab21ecb2b49f37c809f7101106bd12eac5d6" exitCode=0 Dec 04 13:04:28 crc kubenswrapper[4979]: I1204 13:04:28.236557 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"042bf5d7-9b0d-4ee7-84eb-c6209a607717","Type":"ContainerDied","Data":"a5a70153cf74ece77af161689936ab21ecb2b49f37c809f7101106bd12eac5d6"} Dec 04 13:04:28 crc kubenswrapper[4979]: I1204 13:04:28.236729 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-5-default" event={"ID":"042bf5d7-9b0d-4ee7-84eb-c6209a607717","Type":"ContainerStarted","Data":"01f82e8add0d6f14338089e2ae6241ec9dc4d38010d3a2ded24312db3721a3b8"} Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.548331 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.567815 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-5-default_042bf5d7-9b0d-4ee7-84eb-c6209a607717/mariadb-client-5-default/0.log" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.611665 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.618265 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-5-default"] Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.672159 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jb7sj\" (UniqueName: \"kubernetes.io/projected/042bf5d7-9b0d-4ee7-84eb-c6209a607717-kube-api-access-jb7sj\") pod \"042bf5d7-9b0d-4ee7-84eb-c6209a607717\" (UID: \"042bf5d7-9b0d-4ee7-84eb-c6209a607717\") " Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.677629 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/042bf5d7-9b0d-4ee7-84eb-c6209a607717-kube-api-access-jb7sj" (OuterVolumeSpecName: "kube-api-access-jb7sj") pod "042bf5d7-9b0d-4ee7-84eb-c6209a607717" (UID: "042bf5d7-9b0d-4ee7-84eb-c6209a607717"). InnerVolumeSpecName "kube-api-access-jb7sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.741085 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 13:04:29 crc kubenswrapper[4979]: E1204 13:04:29.741503 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="042bf5d7-9b0d-4ee7-84eb-c6209a607717" containerName="mariadb-client-5-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.741529 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="042bf5d7-9b0d-4ee7-84eb-c6209a607717" containerName="mariadb-client-5-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.741709 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="042bf5d7-9b0d-4ee7-84eb-c6209a607717" containerName="mariadb-client-5-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.742320 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.752554 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.774258 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jb7sj\" (UniqueName: \"kubernetes.io/projected/042bf5d7-9b0d-4ee7-84eb-c6209a607717-kube-api-access-jb7sj\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.875691 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-762rc\" (UniqueName: \"kubernetes.io/projected/ed11577c-e4d2-47fd-9331-aa582dfa74fa-kube-api-access-762rc\") pod \"mariadb-client-6-default\" (UID: \"ed11577c-e4d2-47fd-9331-aa582dfa74fa\") " pod="openstack/mariadb-client-6-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.977251 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-762rc\" (UniqueName: \"kubernetes.io/projected/ed11577c-e4d2-47fd-9331-aa582dfa74fa-kube-api-access-762rc\") pod \"mariadb-client-6-default\" (UID: \"ed11577c-e4d2-47fd-9331-aa582dfa74fa\") " pod="openstack/mariadb-client-6-default" Dec 04 13:04:29 crc kubenswrapper[4979]: I1204 13:04:29.993037 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-762rc\" (UniqueName: \"kubernetes.io/projected/ed11577c-e4d2-47fd-9331-aa582dfa74fa-kube-api-access-762rc\") pod \"mariadb-client-6-default\" (UID: \"ed11577c-e4d2-47fd-9331-aa582dfa74fa\") " pod="openstack/mariadb-client-6-default" Dec 04 13:04:30 crc kubenswrapper[4979]: I1204 13:04:30.058463 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 13:04:30 crc kubenswrapper[4979]: I1204 13:04:30.208473 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="042bf5d7-9b0d-4ee7-84eb-c6209a607717" path="/var/lib/kubelet/pods/042bf5d7-9b0d-4ee7-84eb-c6209a607717/volumes" Dec 04 13:04:30 crc kubenswrapper[4979]: I1204 13:04:30.251880 4979 scope.go:117] "RemoveContainer" containerID="a5a70153cf74ece77af161689936ab21ecb2b49f37c809f7101106bd12eac5d6" Dec 04 13:04:30 crc kubenswrapper[4979]: I1204 13:04:30.251921 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-5-default" Dec 04 13:04:30 crc kubenswrapper[4979]: I1204 13:04:30.533086 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 13:04:30 crc kubenswrapper[4979]: W1204 13:04:30.534648 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded11577c_e4d2_47fd_9331_aa582dfa74fa.slice/crio-5e189f6ade54b0e715378b2aea96460d14111f3f7e0279d5a7f5b3c8357c16d0 WatchSource:0}: Error finding container 5e189f6ade54b0e715378b2aea96460d14111f3f7e0279d5a7f5b3c8357c16d0: Status 404 returned error can't find the container with id 5e189f6ade54b0e715378b2aea96460d14111f3f7e0279d5a7f5b3c8357c16d0 Dec 04 13:04:31 crc kubenswrapper[4979]: I1204 13:04:31.268884 4979 generic.go:334] "Generic (PLEG): container finished" podID="ed11577c-e4d2-47fd-9331-aa582dfa74fa" containerID="f6f5067562fb1e83b5940822e0b5dfed2b15298f6121aa32466c36cc1591520c" exitCode=1 Dec 04 13:04:31 crc kubenswrapper[4979]: I1204 13:04:31.268936 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"ed11577c-e4d2-47fd-9331-aa582dfa74fa","Type":"ContainerDied","Data":"f6f5067562fb1e83b5940822e0b5dfed2b15298f6121aa32466c36cc1591520c"} Dec 04 13:04:31 crc kubenswrapper[4979]: I1204 13:04:31.269269 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-6-default" event={"ID":"ed11577c-e4d2-47fd-9331-aa582dfa74fa","Type":"ContainerStarted","Data":"5e189f6ade54b0e715378b2aea96460d14111f3f7e0279d5a7f5b3c8357c16d0"} Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.685352 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.704809 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-6-default_ed11577c-e4d2-47fd-9331-aa582dfa74fa/mariadb-client-6-default/0.log" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.729021 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.734530 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-6-default"] Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.822368 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-762rc\" (UniqueName: \"kubernetes.io/projected/ed11577c-e4d2-47fd-9331-aa582dfa74fa-kube-api-access-762rc\") pod \"ed11577c-e4d2-47fd-9331-aa582dfa74fa\" (UID: \"ed11577c-e4d2-47fd-9331-aa582dfa74fa\") " Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.827945 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed11577c-e4d2-47fd-9331-aa582dfa74fa-kube-api-access-762rc" (OuterVolumeSpecName: "kube-api-access-762rc") pod "ed11577c-e4d2-47fd-9331-aa582dfa74fa" (UID: "ed11577c-e4d2-47fd-9331-aa582dfa74fa"). InnerVolumeSpecName "kube-api-access-762rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.896906 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 13:04:32 crc kubenswrapper[4979]: E1204 13:04:32.897223 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed11577c-e4d2-47fd-9331-aa582dfa74fa" containerName="mariadb-client-6-default" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.897242 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed11577c-e4d2-47fd-9331-aa582dfa74fa" containerName="mariadb-client-6-default" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.897428 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed11577c-e4d2-47fd-9331-aa582dfa74fa" containerName="mariadb-client-6-default" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.897979 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.904117 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 13:04:32 crc kubenswrapper[4979]: I1204 13:04:32.924386 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-762rc\" (UniqueName: \"kubernetes.io/projected/ed11577c-e4d2-47fd-9331-aa582dfa74fa-kube-api-access-762rc\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.026132 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xs96\" (UniqueName: \"kubernetes.io/projected/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b-kube-api-access-2xs96\") pod \"mariadb-client-7-default\" (UID: \"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b\") " pod="openstack/mariadb-client-7-default" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.127477 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xs96\" (UniqueName: \"kubernetes.io/projected/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b-kube-api-access-2xs96\") pod \"mariadb-client-7-default\" (UID: \"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b\") " pod="openstack/mariadb-client-7-default" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.143653 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xs96\" (UniqueName: \"kubernetes.io/projected/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b-kube-api-access-2xs96\") pod \"mariadb-client-7-default\" (UID: \"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b\") " pod="openstack/mariadb-client-7-default" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.214858 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.289351 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e189f6ade54b0e715378b2aea96460d14111f3f7e0279d5a7f5b3c8357c16d0" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.289431 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-6-default" Dec 04 13:04:33 crc kubenswrapper[4979]: I1204 13:04:33.698897 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 13:04:33 crc kubenswrapper[4979]: W1204 13:04:33.706756 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a75b8d3_804c_4c54_a13c_205f5fb0bf0b.slice/crio-0fc9b650c92144de8ce5069e064306b9a895a8685b4572c9193f0929db5f0171 WatchSource:0}: Error finding container 0fc9b650c92144de8ce5069e064306b9a895a8685b4572c9193f0929db5f0171: Status 404 returned error can't find the container with id 0fc9b650c92144de8ce5069e064306b9a895a8685b4572c9193f0929db5f0171 Dec 04 13:04:34 crc kubenswrapper[4979]: I1204 13:04:34.210859 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed11577c-e4d2-47fd-9331-aa582dfa74fa" path="/var/lib/kubelet/pods/ed11577c-e4d2-47fd-9331-aa582dfa74fa/volumes" Dec 04 13:04:34 crc kubenswrapper[4979]: I1204 13:04:34.299657 4979 generic.go:334] "Generic (PLEG): container finished" podID="3a75b8d3-804c-4c54-a13c-205f5fb0bf0b" containerID="05500a45d6535d548e29b1d238633cd21ffe0b03602be46846d695d3dfd99989" exitCode=0 Dec 04 13:04:34 crc kubenswrapper[4979]: I1204 13:04:34.299707 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b","Type":"ContainerDied","Data":"05500a45d6535d548e29b1d238633cd21ffe0b03602be46846d695d3dfd99989"} Dec 04 13:04:34 crc kubenswrapper[4979]: I1204 13:04:34.299742 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-7-default" event={"ID":"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b","Type":"ContainerStarted","Data":"0fc9b650c92144de8ce5069e064306b9a895a8685b4572c9193f0929db5f0171"} Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.637861 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.657347 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-7-default_3a75b8d3-804c-4c54-a13c-205f5fb0bf0b/mariadb-client-7-default/0.log" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.686633 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.692213 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-7-default"] Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.763466 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xs96\" (UniqueName: \"kubernetes.io/projected/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b-kube-api-access-2xs96\") pod \"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b\" (UID: \"3a75b8d3-804c-4c54-a13c-205f5fb0bf0b\") " Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.768516 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b-kube-api-access-2xs96" (OuterVolumeSpecName: "kube-api-access-2xs96") pod "3a75b8d3-804c-4c54-a13c-205f5fb0bf0b" (UID: "3a75b8d3-804c-4c54-a13c-205f5fb0bf0b"). InnerVolumeSpecName "kube-api-access-2xs96". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.826558 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client-2"] Dec 04 13:04:35 crc kubenswrapper[4979]: E1204 13:04:35.827518 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a75b8d3-804c-4c54-a13c-205f5fb0bf0b" containerName="mariadb-client-7-default" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.827542 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a75b8d3-804c-4c54-a13c-205f5fb0bf0b" containerName="mariadb-client-7-default" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.827749 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a75b8d3-804c-4c54-a13c-205f5fb0bf0b" containerName="mariadb-client-7-default" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.828518 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.836495 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.865346 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xs96\" (UniqueName: \"kubernetes.io/projected/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b-kube-api-access-2xs96\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:35 crc kubenswrapper[4979]: I1204 13:04:35.967106 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7h2p6\" (UniqueName: \"kubernetes.io/projected/12c014fc-d8fb-4a47-9b60-68a1c28e7836-kube-api-access-7h2p6\") pod \"mariadb-client-2\" (UID: \"12c014fc-d8fb-4a47-9b60-68a1c28e7836\") " pod="openstack/mariadb-client-2" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.068334 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7h2p6\" (UniqueName: \"kubernetes.io/projected/12c014fc-d8fb-4a47-9b60-68a1c28e7836-kube-api-access-7h2p6\") pod \"mariadb-client-2\" (UID: \"12c014fc-d8fb-4a47-9b60-68a1c28e7836\") " pod="openstack/mariadb-client-2" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.084334 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7h2p6\" (UniqueName: \"kubernetes.io/projected/12c014fc-d8fb-4a47-9b60-68a1c28e7836-kube-api-access-7h2p6\") pod \"mariadb-client-2\" (UID: \"12c014fc-d8fb-4a47-9b60-68a1c28e7836\") " pod="openstack/mariadb-client-2" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.147331 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.208211 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a75b8d3-804c-4c54-a13c-205f5fb0bf0b" path="/var/lib/kubelet/pods/3a75b8d3-804c-4c54-a13c-205f5fb0bf0b/volumes" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.316457 4979 scope.go:117] "RemoveContainer" containerID="05500a45d6535d548e29b1d238633cd21ffe0b03602be46846d695d3dfd99989" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.316478 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-7-default" Dec 04 13:04:36 crc kubenswrapper[4979]: I1204 13:04:36.625739 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 13:04:36 crc kubenswrapper[4979]: W1204 13:04:36.638612 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod12c014fc_d8fb_4a47_9b60_68a1c28e7836.slice/crio-d93df16cf90ef1d291ca6a8ff05c8c517152ebbc4f1ddce1402d602ba55bbf18 WatchSource:0}: Error finding container d93df16cf90ef1d291ca6a8ff05c8c517152ebbc4f1ddce1402d602ba55bbf18: Status 404 returned error can't find the container with id d93df16cf90ef1d291ca6a8ff05c8c517152ebbc4f1ddce1402d602ba55bbf18 Dec 04 13:04:37 crc kubenswrapper[4979]: I1204 13:04:37.325217 4979 generic.go:334] "Generic (PLEG): container finished" podID="12c014fc-d8fb-4a47-9b60-68a1c28e7836" containerID="1af760238713951c6db4cbd83f0171809978ab7502833f77d0b6f4836b22a786" exitCode=0 Dec 04 13:04:37 crc kubenswrapper[4979]: I1204 13:04:37.325358 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"12c014fc-d8fb-4a47-9b60-68a1c28e7836","Type":"ContainerDied","Data":"1af760238713951c6db4cbd83f0171809978ab7502833f77d0b6f4836b22a786"} Dec 04 13:04:37 crc kubenswrapper[4979]: I1204 13:04:37.325548 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client-2" event={"ID":"12c014fc-d8fb-4a47-9b60-68a1c28e7836","Type":"ContainerStarted","Data":"d93df16cf90ef1d291ca6a8ff05c8c517152ebbc4f1ddce1402d602ba55bbf18"} Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.676642 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.695867 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client-2_12c014fc-d8fb-4a47-9b60-68a1c28e7836/mariadb-client-2/0.log" Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.724029 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.730180 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client-2"] Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.807848 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7h2p6\" (UniqueName: \"kubernetes.io/projected/12c014fc-d8fb-4a47-9b60-68a1c28e7836-kube-api-access-7h2p6\") pod \"12c014fc-d8fb-4a47-9b60-68a1c28e7836\" (UID: \"12c014fc-d8fb-4a47-9b60-68a1c28e7836\") " Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.813397 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c014fc-d8fb-4a47-9b60-68a1c28e7836-kube-api-access-7h2p6" (OuterVolumeSpecName: "kube-api-access-7h2p6") pod "12c014fc-d8fb-4a47-9b60-68a1c28e7836" (UID: "12c014fc-d8fb-4a47-9b60-68a1c28e7836"). InnerVolumeSpecName "kube-api-access-7h2p6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:04:38 crc kubenswrapper[4979]: I1204 13:04:38.909279 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7h2p6\" (UniqueName: \"kubernetes.io/projected/12c014fc-d8fb-4a47-9b60-68a1c28e7836-kube-api-access-7h2p6\") on node \"crc\" DevicePath \"\"" Dec 04 13:04:39 crc kubenswrapper[4979]: I1204 13:04:39.342145 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d93df16cf90ef1d291ca6a8ff05c8c517152ebbc4f1ddce1402d602ba55bbf18" Dec 04 13:04:39 crc kubenswrapper[4979]: I1204 13:04:39.342206 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client-2" Dec 04 13:04:40 crc kubenswrapper[4979]: I1204 13:04:40.210486 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c014fc-d8fb-4a47-9b60-68a1c28e7836" path="/var/lib/kubelet/pods/12c014fc-d8fb-4a47-9b60-68a1c28e7836/volumes" Dec 04 13:04:58 crc kubenswrapper[4979]: I1204 13:04:58.040826 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:04:58 crc kubenswrapper[4979]: I1204 13:04:58.041437 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:05:06 crc kubenswrapper[4979]: I1204 13:05:06.076173 4979 scope.go:117] "RemoveContainer" containerID="475a0821eeb4bc52f9dc3d67428ad5c01b618195c83686a1a207e49a24035164" Dec 04 13:05:28 crc kubenswrapper[4979]: I1204 13:05:28.040810 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:05:28 crc kubenswrapper[4979]: I1204 13:05:28.042949 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.040837 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.041460 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.041519 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.042192 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.042251 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" gracePeriod=600 Dec 04 13:05:58 crc kubenswrapper[4979]: E1204 13:05:58.310913 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.947952 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" exitCode=0 Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.948008 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc"} Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.948048 4979 scope.go:117] "RemoveContainer" containerID="2cb5c9544ef934945129415b00a3831fb523499e2744b67a856f00423a013073" Dec 04 13:05:58 crc kubenswrapper[4979]: I1204 13:05:58.948989 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:05:58 crc kubenswrapper[4979]: E1204 13:05:58.949678 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:06:14 crc kubenswrapper[4979]: I1204 13:06:14.199333 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:06:14 crc kubenswrapper[4979]: E1204 13:06:14.200205 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.688361 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-x2v2c"] Dec 04 13:06:22 crc kubenswrapper[4979]: E1204 13:06:22.689097 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c014fc-d8fb-4a47-9b60-68a1c28e7836" containerName="mariadb-client-2" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.689108 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c014fc-d8fb-4a47-9b60-68a1c28e7836" containerName="mariadb-client-2" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.689270 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c014fc-d8fb-4a47-9b60-68a1c28e7836" containerName="mariadb-client-2" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.690333 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.713653 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x2v2c"] Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.774397 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgr69\" (UniqueName: \"kubernetes.io/projected/f09737bf-6522-4e06-b1dc-80673aa05d22-kube-api-access-jgr69\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.774518 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-catalog-content\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.774739 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-utilities\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.876527 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-utilities\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.876598 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgr69\" (UniqueName: \"kubernetes.io/projected/f09737bf-6522-4e06-b1dc-80673aa05d22-kube-api-access-jgr69\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.876660 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-catalog-content\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.877099 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-utilities\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.877319 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-catalog-content\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:22 crc kubenswrapper[4979]: I1204 13:06:22.898073 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgr69\" (UniqueName: \"kubernetes.io/projected/f09737bf-6522-4e06-b1dc-80673aa05d22-kube-api-access-jgr69\") pod \"certified-operators-x2v2c\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:23 crc kubenswrapper[4979]: I1204 13:06:23.007704 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:23 crc kubenswrapper[4979]: I1204 13:06:23.320836 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-x2v2c"] Dec 04 13:06:24 crc kubenswrapper[4979]: I1204 13:06:24.157252 4979 generic.go:334] "Generic (PLEG): container finished" podID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerID="eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e" exitCode=0 Dec 04 13:06:24 crc kubenswrapper[4979]: I1204 13:06:24.157357 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerDied","Data":"eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e"} Dec 04 13:06:24 crc kubenswrapper[4979]: I1204 13:06:24.157720 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerStarted","Data":"5aff0fe179c02fe3250f36b5235f722e62fd912c567d44f42dedf1f8eb8b6bc7"} Dec 04 13:06:24 crc kubenswrapper[4979]: I1204 13:06:24.158997 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:06:25 crc kubenswrapper[4979]: I1204 13:06:25.166741 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerStarted","Data":"47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646"} Dec 04 13:06:26 crc kubenswrapper[4979]: I1204 13:06:26.176010 4979 generic.go:334] "Generic (PLEG): container finished" podID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerID="47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646" exitCode=0 Dec 04 13:06:26 crc kubenswrapper[4979]: I1204 13:06:26.176125 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerDied","Data":"47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646"} Dec 04 13:06:27 crc kubenswrapper[4979]: I1204 13:06:27.184790 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerStarted","Data":"8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23"} Dec 04 13:06:29 crc kubenswrapper[4979]: I1204 13:06:29.199001 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:06:29 crc kubenswrapper[4979]: E1204 13:06:29.199714 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:06:33 crc kubenswrapper[4979]: I1204 13:06:33.008670 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:33 crc kubenswrapper[4979]: I1204 13:06:33.009189 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:33 crc kubenswrapper[4979]: I1204 13:06:33.066183 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:33 crc kubenswrapper[4979]: I1204 13:06:33.093127 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-x2v2c" podStartSLOduration=8.658830707 podStartE2EDuration="11.09310967s" podCreationTimestamp="2025-12-04 13:06:22 +0000 UTC" firstStartedPulling="2025-12-04 13:06:24.158720676 +0000 UTC m=+5008.433016480" lastFinishedPulling="2025-12-04 13:06:26.592999619 +0000 UTC m=+5010.867295443" observedRunningTime="2025-12-04 13:06:27.201735819 +0000 UTC m=+5011.476031633" watchObservedRunningTime="2025-12-04 13:06:33.09310967 +0000 UTC m=+5017.367405474" Dec 04 13:06:33 crc kubenswrapper[4979]: I1204 13:06:33.278026 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:33 crc kubenswrapper[4979]: I1204 13:06:33.329954 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x2v2c"] Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.238727 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-x2v2c" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="registry-server" containerID="cri-o://8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23" gracePeriod=2 Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.626603 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.665039 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-catalog-content\") pod \"f09737bf-6522-4e06-b1dc-80673aa05d22\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.665114 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-utilities\") pod \"f09737bf-6522-4e06-b1dc-80673aa05d22\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.665156 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgr69\" (UniqueName: \"kubernetes.io/projected/f09737bf-6522-4e06-b1dc-80673aa05d22-kube-api-access-jgr69\") pod \"f09737bf-6522-4e06-b1dc-80673aa05d22\" (UID: \"f09737bf-6522-4e06-b1dc-80673aa05d22\") " Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.666313 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-utilities" (OuterVolumeSpecName: "utilities") pod "f09737bf-6522-4e06-b1dc-80673aa05d22" (UID: "f09737bf-6522-4e06-b1dc-80673aa05d22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.677351 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f09737bf-6522-4e06-b1dc-80673aa05d22-kube-api-access-jgr69" (OuterVolumeSpecName: "kube-api-access-jgr69") pod "f09737bf-6522-4e06-b1dc-80673aa05d22" (UID: "f09737bf-6522-4e06-b1dc-80673aa05d22"). InnerVolumeSpecName "kube-api-access-jgr69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.713597 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f09737bf-6522-4e06-b1dc-80673aa05d22" (UID: "f09737bf-6522-4e06-b1dc-80673aa05d22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.766673 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.766716 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f09737bf-6522-4e06-b1dc-80673aa05d22-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:06:35 crc kubenswrapper[4979]: I1204 13:06:35.766729 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgr69\" (UniqueName: \"kubernetes.io/projected/f09737bf-6522-4e06-b1dc-80673aa05d22-kube-api-access-jgr69\") on node \"crc\" DevicePath \"\"" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.247806 4979 generic.go:334] "Generic (PLEG): container finished" podID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerID="8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23" exitCode=0 Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.247959 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-x2v2c" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.247976 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerDied","Data":"8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23"} Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.248599 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-x2v2c" event={"ID":"f09737bf-6522-4e06-b1dc-80673aa05d22","Type":"ContainerDied","Data":"5aff0fe179c02fe3250f36b5235f722e62fd912c567d44f42dedf1f8eb8b6bc7"} Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.248633 4979 scope.go:117] "RemoveContainer" containerID="8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.267499 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-x2v2c"] Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.269136 4979 scope.go:117] "RemoveContainer" containerID="47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.274038 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-x2v2c"] Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.285363 4979 scope.go:117] "RemoveContainer" containerID="eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.315835 4979 scope.go:117] "RemoveContainer" containerID="8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23" Dec 04 13:06:36 crc kubenswrapper[4979]: E1204 13:06:36.316627 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23\": container with ID starting with 8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23 not found: ID does not exist" containerID="8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.316662 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23"} err="failed to get container status \"8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23\": rpc error: code = NotFound desc = could not find container \"8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23\": container with ID starting with 8967629acca1cab28526920121ead31312e8ef3af66316b8be2bc3e3f3b81c23 not found: ID does not exist" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.316685 4979 scope.go:117] "RemoveContainer" containerID="47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646" Dec 04 13:06:36 crc kubenswrapper[4979]: E1204 13:06:36.317081 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646\": container with ID starting with 47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646 not found: ID does not exist" containerID="47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.317103 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646"} err="failed to get container status \"47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646\": rpc error: code = NotFound desc = could not find container \"47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646\": container with ID starting with 47785ca6530b6c176665e12de93ac2fc870727b437e3e106d1c0cd499660d646 not found: ID does not exist" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.317115 4979 scope.go:117] "RemoveContainer" containerID="eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e" Dec 04 13:06:36 crc kubenswrapper[4979]: E1204 13:06:36.317678 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e\": container with ID starting with eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e not found: ID does not exist" containerID="eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e" Dec 04 13:06:36 crc kubenswrapper[4979]: I1204 13:06:36.317782 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e"} err="failed to get container status \"eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e\": rpc error: code = NotFound desc = could not find container \"eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e\": container with ID starting with eb6cfba2556f57b9953e2feea7d138abe263ccfde7db41cda25338e98c4b618e not found: ID does not exist" Dec 04 13:06:38 crc kubenswrapper[4979]: I1204 13:06:38.207226 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" path="/var/lib/kubelet/pods/f09737bf-6522-4e06-b1dc-80673aa05d22/volumes" Dec 04 13:06:40 crc kubenswrapper[4979]: I1204 13:06:40.199733 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:06:40 crc kubenswrapper[4979]: E1204 13:06:40.200038 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:06:53 crc kubenswrapper[4979]: I1204 13:06:53.199424 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:06:53 crc kubenswrapper[4979]: E1204 13:06:53.200113 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:07:06 crc kubenswrapper[4979]: I1204 13:07:06.203675 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:07:06 crc kubenswrapper[4979]: E1204 13:07:06.204286 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:07:19 crc kubenswrapper[4979]: I1204 13:07:19.199162 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:07:19 crc kubenswrapper[4979]: E1204 13:07:19.200002 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:07:31 crc kubenswrapper[4979]: I1204 13:07:31.199575 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:07:31 crc kubenswrapper[4979]: E1204 13:07:31.200587 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:07:46 crc kubenswrapper[4979]: I1204 13:07:46.203375 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:07:46 crc kubenswrapper[4979]: E1204 13:07:46.204109 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.934314 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-92p64"] Dec 04 13:07:56 crc kubenswrapper[4979]: E1204 13:07:56.935167 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="registry-server" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.935185 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="registry-server" Dec 04 13:07:56 crc kubenswrapper[4979]: E1204 13:07:56.935200 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="extract-utilities" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.935207 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="extract-utilities" Dec 04 13:07:56 crc kubenswrapper[4979]: E1204 13:07:56.935220 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="extract-content" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.935228 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="extract-content" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.935423 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f09737bf-6522-4e06-b1dc-80673aa05d22" containerName="registry-server" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.936565 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.948434 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-92p64"] Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.981796 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlk2k\" (UniqueName: \"kubernetes.io/projected/fc18fedd-ee25-48ed-8a76-3d1a480b719f-kube-api-access-xlk2k\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.981866 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-utilities\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:56 crc kubenswrapper[4979]: I1204 13:07:56.981886 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-catalog-content\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.083447 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlk2k\" (UniqueName: \"kubernetes.io/projected/fc18fedd-ee25-48ed-8a76-3d1a480b719f-kube-api-access-xlk2k\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.083510 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-utilities\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.083537 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-catalog-content\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.084053 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-utilities\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.084085 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-catalog-content\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.103952 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlk2k\" (UniqueName: \"kubernetes.io/projected/fc18fedd-ee25-48ed-8a76-3d1a480b719f-kube-api-access-xlk2k\") pod \"community-operators-92p64\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.256749 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.750466 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-92p64"] Dec 04 13:07:57 crc kubenswrapper[4979]: I1204 13:07:57.874094 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerStarted","Data":"738ca0d11d1e9a63d85a8757804b4d522056884a56b10fea07eef2a6b2b20e31"} Dec 04 13:07:58 crc kubenswrapper[4979]: I1204 13:07:58.883417 4979 generic.go:334] "Generic (PLEG): container finished" podID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerID="95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf" exitCode=0 Dec 04 13:07:58 crc kubenswrapper[4979]: I1204 13:07:58.883494 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerDied","Data":"95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf"} Dec 04 13:07:59 crc kubenswrapper[4979]: I1204 13:07:59.198965 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:07:59 crc kubenswrapper[4979]: E1204 13:07:59.199240 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:07:59 crc kubenswrapper[4979]: I1204 13:07:59.899701 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerStarted","Data":"59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba"} Dec 04 13:08:00 crc kubenswrapper[4979]: I1204 13:08:00.911951 4979 generic.go:334] "Generic (PLEG): container finished" podID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerID="59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba" exitCode=0 Dec 04 13:08:00 crc kubenswrapper[4979]: I1204 13:08:00.912024 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerDied","Data":"59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba"} Dec 04 13:08:01 crc kubenswrapper[4979]: I1204 13:08:01.923808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerStarted","Data":"c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8"} Dec 04 13:08:01 crc kubenswrapper[4979]: I1204 13:08:01.950330 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-92p64" podStartSLOduration=3.491787458 podStartE2EDuration="5.950311495s" podCreationTimestamp="2025-12-04 13:07:56 +0000 UTC" firstStartedPulling="2025-12-04 13:07:58.885882531 +0000 UTC m=+5103.160178355" lastFinishedPulling="2025-12-04 13:08:01.344406598 +0000 UTC m=+5105.618702392" observedRunningTime="2025-12-04 13:08:01.9441267 +0000 UTC m=+5106.218422504" watchObservedRunningTime="2025-12-04 13:08:01.950311495 +0000 UTC m=+5106.224607299" Dec 04 13:08:07 crc kubenswrapper[4979]: I1204 13:08:07.257820 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:08:07 crc kubenswrapper[4979]: I1204 13:08:07.257889 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:08:07 crc kubenswrapper[4979]: I1204 13:08:07.303012 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:08:08 crc kubenswrapper[4979]: I1204 13:08:08.008926 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:08:08 crc kubenswrapper[4979]: I1204 13:08:08.058070 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-92p64"] Dec 04 13:08:09 crc kubenswrapper[4979]: I1204 13:08:09.979760 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-92p64" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="registry-server" containerID="cri-o://c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8" gracePeriod=2 Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.405117 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.486582 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlk2k\" (UniqueName: \"kubernetes.io/projected/fc18fedd-ee25-48ed-8a76-3d1a480b719f-kube-api-access-xlk2k\") pod \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.486700 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-catalog-content\") pod \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.486756 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-utilities\") pod \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\" (UID: \"fc18fedd-ee25-48ed-8a76-3d1a480b719f\") " Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.489319 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-utilities" (OuterVolumeSpecName: "utilities") pod "fc18fedd-ee25-48ed-8a76-3d1a480b719f" (UID: "fc18fedd-ee25-48ed-8a76-3d1a480b719f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.493787 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc18fedd-ee25-48ed-8a76-3d1a480b719f-kube-api-access-xlk2k" (OuterVolumeSpecName: "kube-api-access-xlk2k") pod "fc18fedd-ee25-48ed-8a76-3d1a480b719f" (UID: "fc18fedd-ee25-48ed-8a76-3d1a480b719f"). InnerVolumeSpecName "kube-api-access-xlk2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.546628 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc18fedd-ee25-48ed-8a76-3d1a480b719f" (UID: "fc18fedd-ee25-48ed-8a76-3d1a480b719f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.588027 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlk2k\" (UniqueName: \"kubernetes.io/projected/fc18fedd-ee25-48ed-8a76-3d1a480b719f-kube-api-access-xlk2k\") on node \"crc\" DevicePath \"\"" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.588431 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.588445 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc18fedd-ee25-48ed-8a76-3d1a480b719f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.990367 4979 generic.go:334] "Generic (PLEG): container finished" podID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerID="c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8" exitCode=0 Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.990420 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerDied","Data":"c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8"} Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.990464 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92p64" event={"ID":"fc18fedd-ee25-48ed-8a76-3d1a480b719f","Type":"ContainerDied","Data":"738ca0d11d1e9a63d85a8757804b4d522056884a56b10fea07eef2a6b2b20e31"} Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.990484 4979 scope.go:117] "RemoveContainer" containerID="c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8" Dec 04 13:08:10 crc kubenswrapper[4979]: I1204 13:08:10.990704 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92p64" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.011339 4979 scope.go:117] "RemoveContainer" containerID="59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.029268 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-92p64"] Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.040409 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-92p64"] Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.058633 4979 scope.go:117] "RemoveContainer" containerID="95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.084750 4979 scope.go:117] "RemoveContainer" containerID="c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8" Dec 04 13:08:11 crc kubenswrapper[4979]: E1204 13:08:11.085320 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8\": container with ID starting with c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8 not found: ID does not exist" containerID="c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.085363 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8"} err="failed to get container status \"c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8\": rpc error: code = NotFound desc = could not find container \"c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8\": container with ID starting with c8e6f18ed326aa83f0f389ddb861125d3a0a33330a3b0c8aa09aec23128cdec8 not found: ID does not exist" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.085390 4979 scope.go:117] "RemoveContainer" containerID="59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba" Dec 04 13:08:11 crc kubenswrapper[4979]: E1204 13:08:11.085791 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba\": container with ID starting with 59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba not found: ID does not exist" containerID="59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.085817 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba"} err="failed to get container status \"59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba\": rpc error: code = NotFound desc = could not find container \"59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba\": container with ID starting with 59340dc6619a4507ad4997dc2a128657049aa6b55ee5d816bd55a356f9fe57ba not found: ID does not exist" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.085834 4979 scope.go:117] "RemoveContainer" containerID="95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf" Dec 04 13:08:11 crc kubenswrapper[4979]: E1204 13:08:11.086243 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf\": container with ID starting with 95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf not found: ID does not exist" containerID="95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf" Dec 04 13:08:11 crc kubenswrapper[4979]: I1204 13:08:11.086267 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf"} err="failed to get container status \"95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf\": rpc error: code = NotFound desc = could not find container \"95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf\": container with ID starting with 95313b93458f5f29a37480b28652c23aff8fd725c70e6cbfdd897d590697fdaf not found: ID does not exist" Dec 04 13:08:12 crc kubenswrapper[4979]: I1204 13:08:12.208198 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" path="/var/lib/kubelet/pods/fc18fedd-ee25-48ed-8a76-3d1a480b719f/volumes" Dec 04 13:08:14 crc kubenswrapper[4979]: I1204 13:08:14.200478 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:08:14 crc kubenswrapper[4979]: E1204 13:08:14.201050 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:08:25 crc kubenswrapper[4979]: I1204 13:08:25.199223 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:08:25 crc kubenswrapper[4979]: E1204 13:08:25.200116 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:08:38 crc kubenswrapper[4979]: I1204 13:08:38.206854 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:08:38 crc kubenswrapper[4979]: E1204 13:08:38.208423 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:08:53 crc kubenswrapper[4979]: I1204 13:08:53.199474 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:08:53 crc kubenswrapper[4979]: E1204 13:08:53.200137 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.297230 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-645tm"] Dec 04 13:09:02 crc kubenswrapper[4979]: E1204 13:09:02.298276 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="extract-utilities" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.298292 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="extract-utilities" Dec 04 13:09:02 crc kubenswrapper[4979]: E1204 13:09:02.298327 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="registry-server" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.298333 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="registry-server" Dec 04 13:09:02 crc kubenswrapper[4979]: E1204 13:09:02.298342 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="extract-content" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.298349 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="extract-content" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.298488 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc18fedd-ee25-48ed-8a76-3d1a480b719f" containerName="registry-server" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.299781 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.312530 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-645tm"] Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.360869 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-catalog-content\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.361012 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-utilities\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.361081 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx5pm\" (UniqueName: \"kubernetes.io/projected/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-kube-api-access-bx5pm\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.463106 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-utilities\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.463205 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx5pm\" (UniqueName: \"kubernetes.io/projected/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-kube-api-access-bx5pm\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.463290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-catalog-content\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.463816 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-utilities\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.463839 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-catalog-content\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.484227 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx5pm\" (UniqueName: \"kubernetes.io/projected/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-kube-api-access-bx5pm\") pod \"redhat-operators-645tm\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:02 crc kubenswrapper[4979]: I1204 13:09:02.621213 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:03 crc kubenswrapper[4979]: I1204 13:09:03.056655 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-645tm"] Dec 04 13:09:03 crc kubenswrapper[4979]: I1204 13:09:03.406584 4979 generic.go:334] "Generic (PLEG): container finished" podID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerID="87a9fe26526fbd613efb06640a199d06e620b37038434f95d51c958404fe0d18" exitCode=0 Dec 04 13:09:03 crc kubenswrapper[4979]: I1204 13:09:03.406644 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerDied","Data":"87a9fe26526fbd613efb06640a199d06e620b37038434f95d51c958404fe0d18"} Dec 04 13:09:03 crc kubenswrapper[4979]: I1204 13:09:03.406953 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerStarted","Data":"b605401d399e795c3272e3da280d2ab20551798b143e09429313ddb3586e377e"} Dec 04 13:09:04 crc kubenswrapper[4979]: I1204 13:09:04.417712 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerStarted","Data":"20e566bf7c2bee0e5900e8612dfb7988368c08749f3beda26ae39714fa35e9d9"} Dec 04 13:09:05 crc kubenswrapper[4979]: I1204 13:09:05.428413 4979 generic.go:334] "Generic (PLEG): container finished" podID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerID="20e566bf7c2bee0e5900e8612dfb7988368c08749f3beda26ae39714fa35e9d9" exitCode=0 Dec 04 13:09:05 crc kubenswrapper[4979]: I1204 13:09:05.428470 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerDied","Data":"20e566bf7c2bee0e5900e8612dfb7988368c08749f3beda26ae39714fa35e9d9"} Dec 04 13:09:06 crc kubenswrapper[4979]: I1204 13:09:06.204023 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:09:06 crc kubenswrapper[4979]: E1204 13:09:06.204519 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:09:06 crc kubenswrapper[4979]: I1204 13:09:06.444640 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerStarted","Data":"2714ab2ba0c76625e12a736d6d5441cda00aff67a9d4bf5c6121b173a1ba8309"} Dec 04 13:09:06 crc kubenswrapper[4979]: I1204 13:09:06.462177 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-645tm" podStartSLOduration=2.044487767 podStartE2EDuration="4.462158867s" podCreationTimestamp="2025-12-04 13:09:02 +0000 UTC" firstStartedPulling="2025-12-04 13:09:03.408103279 +0000 UTC m=+5167.682399083" lastFinishedPulling="2025-12-04 13:09:05.825774379 +0000 UTC m=+5170.100070183" observedRunningTime="2025-12-04 13:09:06.461637733 +0000 UTC m=+5170.735933537" watchObservedRunningTime="2025-12-04 13:09:06.462158867 +0000 UTC m=+5170.736454671" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.417170 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.418514 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.421192 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-wx59x" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.427961 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.554430 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.554512 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k48nh\" (UniqueName: \"kubernetes.io/projected/bfae82d7-d730-474e-acb8-e0b5ef2630ab-kube-api-access-k48nh\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.656329 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.656409 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k48nh\" (UniqueName: \"kubernetes.io/projected/bfae82d7-d730-474e-acb8-e0b5ef2630ab-kube-api-access-k48nh\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.659560 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.659635 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/12d5bfacb7b3c4d423e17bf542b77b82b1a097b022754f489e3c37fbd245b17a/globalmount\"" pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.678367 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k48nh\" (UniqueName: \"kubernetes.io/projected/bfae82d7-d730-474e-acb8-e0b5ef2630ab-kube-api-access-k48nh\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.686845 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") pod \"mariadb-copy-data\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " pod="openstack/mariadb-copy-data" Dec 04 13:09:08 crc kubenswrapper[4979]: I1204 13:09:08.747732 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 13:09:09 crc kubenswrapper[4979]: I1204 13:09:09.242958 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 13:09:09 crc kubenswrapper[4979]: W1204 13:09:09.247101 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbfae82d7_d730_474e_acb8_e0b5ef2630ab.slice/crio-393b2e9566b3ae49806e40aa42874273f225c9769d0eb80eb9e0c243a9f252cd WatchSource:0}: Error finding container 393b2e9566b3ae49806e40aa42874273f225c9769d0eb80eb9e0c243a9f252cd: Status 404 returned error can't find the container with id 393b2e9566b3ae49806e40aa42874273f225c9769d0eb80eb9e0c243a9f252cd Dec 04 13:09:09 crc kubenswrapper[4979]: I1204 13:09:09.470257 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"bfae82d7-d730-474e-acb8-e0b5ef2630ab","Type":"ContainerStarted","Data":"393b2e9566b3ae49806e40aa42874273f225c9769d0eb80eb9e0c243a9f252cd"} Dec 04 13:09:10 crc kubenswrapper[4979]: I1204 13:09:10.478995 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"bfae82d7-d730-474e-acb8-e0b5ef2630ab","Type":"ContainerStarted","Data":"9c0acd65ffc63a662e9cf3545724e6de8d636eab7f51c3e02cc5072fe5758ea1"} Dec 04 13:09:10 crc kubenswrapper[4979]: I1204 13:09:10.497768 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-copy-data" podStartSLOduration=3.497750527 podStartE2EDuration="3.497750527s" podCreationTimestamp="2025-12-04 13:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:09:10.493648348 +0000 UTC m=+5174.767944152" watchObservedRunningTime="2025-12-04 13:09:10.497750527 +0000 UTC m=+5174.772046321" Dec 04 13:09:12 crc kubenswrapper[4979]: I1204 13:09:12.623554 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:12 crc kubenswrapper[4979]: I1204 13:09:12.624491 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:12 crc kubenswrapper[4979]: I1204 13:09:12.671117 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.374500 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.375635 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.380750 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.529805 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwfl\" (UniqueName: \"kubernetes.io/projected/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6-kube-api-access-wpwfl\") pod \"mariadb-client\" (UID: \"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6\") " pod="openstack/mariadb-client" Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.549002 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.594706 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-645tm"] Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.631074 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwfl\" (UniqueName: \"kubernetes.io/projected/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6-kube-api-access-wpwfl\") pod \"mariadb-client\" (UID: \"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6\") " pod="openstack/mariadb-client" Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.653806 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwfl\" (UniqueName: \"kubernetes.io/projected/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6-kube-api-access-wpwfl\") pod \"mariadb-client\" (UID: \"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6\") " pod="openstack/mariadb-client" Dec 04 13:09:13 crc kubenswrapper[4979]: I1204 13:09:13.695052 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:14 crc kubenswrapper[4979]: I1204 13:09:14.103933 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:14 crc kubenswrapper[4979]: W1204 13:09:14.110665 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0d57727_7ce5_4ea3_a8ec_ebf3ae23e9d6.slice/crio-afc224dac8939df2a4a76a5161866a5d5b2fe56c356be60abfed0fe09e7fc1e0 WatchSource:0}: Error finding container afc224dac8939df2a4a76a5161866a5d5b2fe56c356be60abfed0fe09e7fc1e0: Status 404 returned error can't find the container with id afc224dac8939df2a4a76a5161866a5d5b2fe56c356be60abfed0fe09e7fc1e0 Dec 04 13:09:14 crc kubenswrapper[4979]: I1204 13:09:14.510075 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6","Type":"ContainerStarted","Data":"afc224dac8939df2a4a76a5161866a5d5b2fe56c356be60abfed0fe09e7fc1e0"} Dec 04 13:09:15 crc kubenswrapper[4979]: I1204 13:09:15.516199 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-645tm" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="registry-server" containerID="cri-o://2714ab2ba0c76625e12a736d6d5441cda00aff67a9d4bf5c6121b173a1ba8309" gracePeriod=2 Dec 04 13:09:16 crc kubenswrapper[4979]: I1204 13:09:16.524608 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6","Type":"ContainerStarted","Data":"25d1b90d97e23f949dff8d8ff6b8dee2c636d6afc3b6b8b9317f129690592856"} Dec 04 13:09:17 crc kubenswrapper[4979]: I1204 13:09:17.199833 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:09:17 crc kubenswrapper[4979]: E1204 13:09:17.200084 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:09:17 crc kubenswrapper[4979]: I1204 13:09:17.552918 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mariadb-client" podStartSLOduration=4.552897509 podStartE2EDuration="4.552897509s" podCreationTimestamp="2025-12-04 13:09:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:09:17.547494315 +0000 UTC m=+5181.821790139" watchObservedRunningTime="2025-12-04 13:09:17.552897509 +0000 UTC m=+5181.827193313" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.126971 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6/mariadb-client/0.log" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.541364 4979 generic.go:334] "Generic (PLEG): container finished" podID="e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6" containerID="25d1b90d97e23f949dff8d8ff6b8dee2c636d6afc3b6b8b9317f129690592856" exitCode=0 Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.541428 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6","Type":"ContainerDied","Data":"25d1b90d97e23f949dff8d8ff6b8dee2c636d6afc3b6b8b9317f129690592856"} Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.543788 4979 generic.go:334] "Generic (PLEG): container finished" podID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerID="2714ab2ba0c76625e12a736d6d5441cda00aff67a9d4bf5c6121b173a1ba8309" exitCode=0 Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.543870 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerDied","Data":"2714ab2ba0c76625e12a736d6d5441cda00aff67a9d4bf5c6121b173a1ba8309"} Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.613835 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.710230 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx5pm\" (UniqueName: \"kubernetes.io/projected/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-kube-api-access-bx5pm\") pod \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.710314 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-catalog-content\") pod \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.710439 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-utilities\") pod \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\" (UID: \"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313\") " Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.711491 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-utilities" (OuterVolumeSpecName: "utilities") pod "9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" (UID: "9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.718383 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-kube-api-access-bx5pm" (OuterVolumeSpecName: "kube-api-access-bx5pm") pod "9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" (UID: "9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313"). InnerVolumeSpecName "kube-api-access-bx5pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.813029 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx5pm\" (UniqueName: \"kubernetes.io/projected/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-kube-api-access-bx5pm\") on node \"crc\" DevicePath \"\"" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.813086 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.829866 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" (UID: "9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:09:18 crc kubenswrapper[4979]: I1204 13:09:18.914534 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.556031 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-645tm" Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.556132 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-645tm" event={"ID":"9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313","Type":"ContainerDied","Data":"b605401d399e795c3272e3da280d2ab20551798b143e09429313ddb3586e377e"} Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.556184 4979 scope.go:117] "RemoveContainer" containerID="2714ab2ba0c76625e12a736d6d5441cda00aff67a9d4bf5c6121b173a1ba8309" Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.575717 4979 scope.go:117] "RemoveContainer" containerID="20e566bf7c2bee0e5900e8612dfb7988368c08749f3beda26ae39714fa35e9d9" Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.593523 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-645tm"] Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.601491 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-645tm"] Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.622204 4979 scope.go:117] "RemoveContainer" containerID="87a9fe26526fbd613efb06640a199d06e620b37038434f95d51c958404fe0d18" Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.873871 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.913827 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:19 crc kubenswrapper[4979]: I1204 13:09:19.921287 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.029342 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpwfl\" (UniqueName: \"kubernetes.io/projected/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6-kube-api-access-wpwfl\") pod \"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6\" (UID: \"e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6\") " Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.038802 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6-kube-api-access-wpwfl" (OuterVolumeSpecName: "kube-api-access-wpwfl") pod "e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6" (UID: "e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6"). InnerVolumeSpecName "kube-api-access-wpwfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052095 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:20 crc kubenswrapper[4979]: E1204 13:09:20.052563 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="extract-utilities" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052585 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="extract-utilities" Dec 04 13:09:20 crc kubenswrapper[4979]: E1204 13:09:20.052597 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6" containerName="mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052606 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6" containerName="mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: E1204 13:09:20.052635 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="extract-content" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052643 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="extract-content" Dec 04 13:09:20 crc kubenswrapper[4979]: E1204 13:09:20.052664 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="registry-server" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052673 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="registry-server" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052834 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6" containerName="mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.052861 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" containerName="registry-server" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.053527 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.058495 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.131697 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpwfl\" (UniqueName: \"kubernetes.io/projected/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6-kube-api-access-wpwfl\") on node \"crc\" DevicePath \"\"" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.212967 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313" path="/var/lib/kubelet/pods/9cf11a91-a0d6-4d5d-b9fd-f7d7f710e313/volumes" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.214526 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6" path="/var/lib/kubelet/pods/e0d57727-7ce5-4ea3-a8ec-ebf3ae23e9d6/volumes" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.233285 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85xj6\" (UniqueName: \"kubernetes.io/projected/9e9f0b93-ff36-41de-88af-97fbcf4cdda6-kube-api-access-85xj6\") pod \"mariadb-client\" (UID: \"9e9f0b93-ff36-41de-88af-97fbcf4cdda6\") " pod="openstack/mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.335245 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85xj6\" (UniqueName: \"kubernetes.io/projected/9e9f0b93-ff36-41de-88af-97fbcf4cdda6-kube-api-access-85xj6\") pod \"mariadb-client\" (UID: \"9e9f0b93-ff36-41de-88af-97fbcf4cdda6\") " pod="openstack/mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.354277 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85xj6\" (UniqueName: \"kubernetes.io/projected/9e9f0b93-ff36-41de-88af-97fbcf4cdda6-kube-api-access-85xj6\") pod \"mariadb-client\" (UID: \"9e9f0b93-ff36-41de-88af-97fbcf4cdda6\") " pod="openstack/mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.396252 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.564752 4979 scope.go:117] "RemoveContainer" containerID="25d1b90d97e23f949dff8d8ff6b8dee2c636d6afc3b6b8b9317f129690592856" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.564898 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:20 crc kubenswrapper[4979]: I1204 13:09:20.811593 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:20 crc kubenswrapper[4979]: W1204 13:09:20.816415 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9e9f0b93_ff36_41de_88af_97fbcf4cdda6.slice/crio-179310ef50def28697fd8813bb796509f929f1fd1fade9f4d715beab2500b702 WatchSource:0}: Error finding container 179310ef50def28697fd8813bb796509f929f1fd1fade9f4d715beab2500b702: Status 404 returned error can't find the container with id 179310ef50def28697fd8813bb796509f929f1fd1fade9f4d715beab2500b702 Dec 04 13:09:21 crc kubenswrapper[4979]: I1204 13:09:21.576947 4979 generic.go:334] "Generic (PLEG): container finished" podID="9e9f0b93-ff36-41de-88af-97fbcf4cdda6" containerID="bcd8476674daa4aad1246b52d39b39d1a9ab9763c05832179368c1b495b7aed0" exitCode=0 Dec 04 13:09:21 crc kubenswrapper[4979]: I1204 13:09:21.577025 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"9e9f0b93-ff36-41de-88af-97fbcf4cdda6","Type":"ContainerDied","Data":"bcd8476674daa4aad1246b52d39b39d1a9ab9763c05832179368c1b495b7aed0"} Dec 04 13:09:21 crc kubenswrapper[4979]: I1204 13:09:21.577360 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-client" event={"ID":"9e9f0b93-ff36-41de-88af-97fbcf4cdda6","Type":"ContainerStarted","Data":"179310ef50def28697fd8813bb796509f929f1fd1fade9f4d715beab2500b702"} Dec 04 13:09:22 crc kubenswrapper[4979]: I1204 13:09:22.862039 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:22 crc kubenswrapper[4979]: I1204 13:09:22.881816 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mariadb-client_9e9f0b93-ff36-41de-88af-97fbcf4cdda6/mariadb-client/0.log" Dec 04 13:09:22 crc kubenswrapper[4979]: I1204 13:09:22.908176 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:22 crc kubenswrapper[4979]: I1204 13:09:22.913523 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-client"] Dec 04 13:09:22 crc kubenswrapper[4979]: I1204 13:09:22.976462 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85xj6\" (UniqueName: \"kubernetes.io/projected/9e9f0b93-ff36-41de-88af-97fbcf4cdda6-kube-api-access-85xj6\") pod \"9e9f0b93-ff36-41de-88af-97fbcf4cdda6\" (UID: \"9e9f0b93-ff36-41de-88af-97fbcf4cdda6\") " Dec 04 13:09:22 crc kubenswrapper[4979]: I1204 13:09:22.981660 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9f0b93-ff36-41de-88af-97fbcf4cdda6-kube-api-access-85xj6" (OuterVolumeSpecName: "kube-api-access-85xj6") pod "9e9f0b93-ff36-41de-88af-97fbcf4cdda6" (UID: "9e9f0b93-ff36-41de-88af-97fbcf4cdda6"). InnerVolumeSpecName "kube-api-access-85xj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:09:23 crc kubenswrapper[4979]: I1204 13:09:23.078516 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85xj6\" (UniqueName: \"kubernetes.io/projected/9e9f0b93-ff36-41de-88af-97fbcf4cdda6-kube-api-access-85xj6\") on node \"crc\" DevicePath \"\"" Dec 04 13:09:23 crc kubenswrapper[4979]: I1204 13:09:23.594343 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="179310ef50def28697fd8813bb796509f929f1fd1fade9f4d715beab2500b702" Dec 04 13:09:23 crc kubenswrapper[4979]: I1204 13:09:23.594425 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-client" Dec 04 13:09:24 crc kubenswrapper[4979]: I1204 13:09:24.206891 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9f0b93-ff36-41de-88af-97fbcf4cdda6" path="/var/lib/kubelet/pods/9e9f0b93-ff36-41de-88af-97fbcf4cdda6/volumes" Dec 04 13:09:31 crc kubenswrapper[4979]: I1204 13:09:31.199588 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:09:31 crc kubenswrapper[4979]: E1204 13:09:31.200374 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:09:46 crc kubenswrapper[4979]: I1204 13:09:46.205672 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:09:46 crc kubenswrapper[4979]: E1204 13:09:46.206448 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:09:59 crc kubenswrapper[4979]: I1204 13:09:59.199442 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:09:59 crc kubenswrapper[4979]: E1204 13:09:59.200292 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:10:11 crc kubenswrapper[4979]: I1204 13:10:11.198875 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:10:11 crc kubenswrapper[4979]: E1204 13:10:11.199773 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.872079 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 13:10:14 crc kubenswrapper[4979]: E1204 13:10:14.873095 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e9f0b93-ff36-41de-88af-97fbcf4cdda6" containerName="mariadb-client" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.873112 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e9f0b93-ff36-41de-88af-97fbcf4cdda6" containerName="mariadb-client" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.873314 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e9f0b93-ff36-41de-88af-97fbcf4cdda6" containerName="mariadb-client" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.874336 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.876322 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-6jm9j" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.878439 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.878695 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.894285 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.904368 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.906033 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.922537 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.924736 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.944484 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 04 13:10:14 crc kubenswrapper[4979]: I1204 13:10:14.960887 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072637 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-94d98d40-8520-4def-81f4-f3a0c594f410\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94d98d40-8520-4def-81f4-f3a0c594f410\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072699 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd92753a-87a9-47ca-ae6b-add04cbf46d3-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072774 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6daaa5ce-1809-4b88-bec5-90204cbc1491-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072863 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd92753a-87a9-47ca-ae6b-add04cbf46d3-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072897 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072942 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6daaa5ce-1809-4b88-bec5-90204cbc1491-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.072986 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6daaa5ce-1809-4b88-bec5-90204cbc1491-config\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073014 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn95w\" (UniqueName: \"kubernetes.io/projected/6daaa5ce-1809-4b88-bec5-90204cbc1491-kube-api-access-gn95w\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073046 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd92753a-87a9-47ca-ae6b-add04cbf46d3-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073129 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d9841cd-dc04-4f50-8436-3d020df5f2f2-config\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073199 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d9841cd-dc04-4f50-8436-3d020df5f2f2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073250 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr287\" (UniqueName: \"kubernetes.io/projected/dd92753a-87a9-47ca-ae6b-add04cbf46d3-kube-api-access-dr287\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073348 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073383 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9841cd-dc04-4f50-8436-3d020df5f2f2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073409 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd92753a-87a9-47ca-ae6b-add04cbf46d3-config\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073459 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9841cd-dc04-4f50-8436-3d020df5f2f2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073479 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcbf4\" (UniqueName: \"kubernetes.io/projected/4d9841cd-dc04-4f50-8436-3d020df5f2f2-kube-api-access-xcbf4\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.073517 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6daaa5ce-1809-4b88-bec5-90204cbc1491-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.077806 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.079029 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.080638 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-xkqcm" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.081152 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.083078 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.092019 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.104084 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.105668 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.121247 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.123545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.131474 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.143455 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174707 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c0f621-2918-4193-9420-780b3ecbd5e5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174763 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd92753a-87a9-47ca-ae6b-add04cbf46d3-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174787 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174807 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6daaa5ce-1809-4b88-bec5-90204cbc1491-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174829 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6daaa5ce-1809-4b88-bec5-90204cbc1491-config\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174852 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn95w\" (UniqueName: \"kubernetes.io/projected/6daaa5ce-1809-4b88-bec5-90204cbc1491-kube-api-access-gn95w\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174875 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd92753a-87a9-47ca-ae6b-add04cbf46d3-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d9841cd-dc04-4f50-8436-3d020df5f2f2-config\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174914 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d9841cd-dc04-4f50-8436-3d020df5f2f2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174935 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr287\" (UniqueName: \"kubernetes.io/projected/dd92753a-87a9-47ca-ae6b-add04cbf46d3-kube-api-access-dr287\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174959 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.174981 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62c0f621-2918-4193-9420-780b3ecbd5e5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175004 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9841cd-dc04-4f50-8436-3d020df5f2f2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175024 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd92753a-87a9-47ca-ae6b-add04cbf46d3-config\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175051 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9841cd-dc04-4f50-8436-3d020df5f2f2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175071 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62c0f621-2918-4193-9420-780b3ecbd5e5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175091 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcbf4\" (UniqueName: \"kubernetes.io/projected/4d9841cd-dc04-4f50-8436-3d020df5f2f2-kube-api-access-xcbf4\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175114 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq22t\" (UniqueName: \"kubernetes.io/projected/62c0f621-2918-4193-9420-780b3ecbd5e5-kube-api-access-fq22t\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175136 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6daaa5ce-1809-4b88-bec5-90204cbc1491-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175163 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd92753a-87a9-47ca-ae6b-add04cbf46d3-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175182 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ba8d8516-ad5e-4245-a877-48487d19e088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba8d8516-ad5e-4245-a877-48487d19e088\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175199 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-94d98d40-8520-4def-81f4-f3a0c594f410\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94d98d40-8520-4def-81f4-f3a0c594f410\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175237 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0f621-2918-4193-9420-780b3ecbd5e5-config\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.175269 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6daaa5ce-1809-4b88-bec5-90204cbc1491-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.177109 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6daaa5ce-1809-4b88-bec5-90204cbc1491-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.177917 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dd92753a-87a9-47ca-ae6b-add04cbf46d3-scripts\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.177937 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6daaa5ce-1809-4b88-bec5-90204cbc1491-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.178363 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6daaa5ce-1809-4b88-bec5-90204cbc1491-config\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.178709 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4d9841cd-dc04-4f50-8436-3d020df5f2f2-ovsdb-rundir\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.179734 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd92753a-87a9-47ca-ae6b-add04cbf46d3-config\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.179742 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d9841cd-dc04-4f50-8436-3d020df5f2f2-scripts\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.180223 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/dd92753a-87a9-47ca-ae6b-add04cbf46d3-ovsdb-rundir\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.181513 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.181547 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0c0ba6496748eb95bcb67c4867d161a9c48be39cf90803d9883d9c36b325bb00/globalmount\"" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.181710 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.181753 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-94d98d40-8520-4def-81f4-f3a0c594f410\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94d98d40-8520-4def-81f4-f3a0c594f410\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ef5db64f9aa542c3647870613aee160fd691d22f924e66a273213441227f084b/globalmount\"" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.182471 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.182509 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a971bd77e37230990e851789f0bbac20c2f9e80190461bdb7a3138ee66bb3f52/globalmount\"" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.182800 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d9841cd-dc04-4f50-8436-3d020df5f2f2-config\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.191099 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd92753a-87a9-47ca-ae6b-add04cbf46d3-combined-ca-bundle\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.194869 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9841cd-dc04-4f50-8436-3d020df5f2f2-combined-ca-bundle\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.195609 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6daaa5ce-1809-4b88-bec5-90204cbc1491-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.196028 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr287\" (UniqueName: \"kubernetes.io/projected/dd92753a-87a9-47ca-ae6b-add04cbf46d3-kube-api-access-dr287\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.197651 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn95w\" (UniqueName: \"kubernetes.io/projected/6daaa5ce-1809-4b88-bec5-90204cbc1491-kube-api-access-gn95w\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.198150 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcbf4\" (UniqueName: \"kubernetes.io/projected/4d9841cd-dc04-4f50-8436-3d020df5f2f2-kube-api-access-xcbf4\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.214332 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-94d98d40-8520-4def-81f4-f3a0c594f410\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-94d98d40-8520-4def-81f4-f3a0c594f410\") pod \"ovsdbserver-sb-0\" (UID: \"6daaa5ce-1809-4b88-bec5-90204cbc1491\") " pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.218718 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e867aaa6-ab6d-4645-80ce-0a4a5d400d16\") pod \"ovsdbserver-sb-1\" (UID: \"4d9841cd-dc04-4f50-8436-3d020df5f2f2\") " pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.223566 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-5806703a-6f7d-4864-9cc4-36714ad01d68\") pod \"ovsdbserver-sb-2\" (UID: \"dd92753a-87a9-47ca-ae6b-add04cbf46d3\") " pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.240197 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.260144 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.276881 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62c0f621-2918-4193-9420-780b3ecbd5e5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.276960 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277001 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62c0f621-2918-4193-9420-780b3ecbd5e5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq22t\" (UniqueName: \"kubernetes.io/projected/62c0f621-2918-4193-9420-780b3ecbd5e5-kube-api-access-fq22t\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277054 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277082 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e0f5920-9f22-4288-b09f-be3e565d88df-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277108 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ba8d8516-ad5e-4245-a877-48487d19e088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba8d8516-ad5e-4245-a877-48487d19e088\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277166 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277220 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9jln\" (UniqueName: \"kubernetes.io/projected/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-kube-api-access-j9jln\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277246 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0f621-2918-4193-9420-780b3ecbd5e5-config\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277274 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghdff\" (UniqueName: \"kubernetes.io/projected/2e0f5920-9f22-4288-b09f-be3e565d88df-kube-api-access-ghdff\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277338 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277362 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e0f5920-9f22-4288-b09f-be3e565d88df-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277393 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c0f621-2918-4193-9420-780b3ecbd5e5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277416 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f5920-9f22-4288-b09f-be3e565d88df-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277459 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e0f5920-9f22-4288-b09f-be3e565d88df-config\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277483 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.277544 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-config\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.278286 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/62c0f621-2918-4193-9420-780b3ecbd5e5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.282615 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c0f621-2918-4193-9420-780b3ecbd5e5-config\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.284784 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/62c0f621-2918-4193-9420-780b3ecbd5e5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.289484 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c0f621-2918-4193-9420-780b3ecbd5e5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.289760 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.289807 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ba8d8516-ad5e-4245-a877-48487d19e088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba8d8516-ad5e-4245-a877-48487d19e088\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/2895750db0caab3626a6bccb905a8f9f1a11128c8c7b33b406c92ed5846ca350/globalmount\"" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.299278 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq22t\" (UniqueName: \"kubernetes.io/projected/62c0f621-2918-4193-9420-780b3ecbd5e5-kube-api-access-fq22t\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.322406 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ba8d8516-ad5e-4245-a877-48487d19e088\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ba8d8516-ad5e-4245-a877-48487d19e088\") pod \"ovsdbserver-nb-0\" (UID: \"62c0f621-2918-4193-9420-780b3ecbd5e5\") " pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379517 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-config\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379609 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379689 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379740 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e0f5920-9f22-4288-b09f-be3e565d88df-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379770 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379837 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9jln\" (UniqueName: \"kubernetes.io/projected/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-kube-api-access-j9jln\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379866 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghdff\" (UniqueName: \"kubernetes.io/projected/2e0f5920-9f22-4288-b09f-be3e565d88df-kube-api-access-ghdff\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379922 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.379948 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e0f5920-9f22-4288-b09f-be3e565d88df-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.380000 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f5920-9f22-4288-b09f-be3e565d88df-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.380080 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e0f5920-9f22-4288-b09f-be3e565d88df-config\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.380100 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.381072 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/2e0f5920-9f22-4288-b09f-be3e565d88df-ovsdb-rundir\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.381369 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-config\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.381562 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-ovsdb-rundir\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.382741 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2e0f5920-9f22-4288-b09f-be3e565d88df-config\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.384720 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-scripts\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.385254 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e0f5920-9f22-4288-b09f-be3e565d88df-scripts\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.385815 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e0f5920-9f22-4288-b09f-be3e565d88df-combined-ca-bundle\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.386831 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.386863 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/fa54a18e5e5872749d888589398fee3908d2047ccbf24f74a0f4bfa31c852e64/globalmount\"" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.386991 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.387040 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c57fc7b5bcc4098297b3024672baac460ecc2647a31f19dba23e0e012cdb0c76/globalmount\"" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.388834 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-combined-ca-bundle\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.398056 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghdff\" (UniqueName: \"kubernetes.io/projected/2e0f5920-9f22-4288-b09f-be3e565d88df-kube-api-access-ghdff\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.399563 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9jln\" (UniqueName: \"kubernetes.io/projected/d3ac5ba2-798d-4101-ad63-14cf9cfd47b4-kube-api-access-j9jln\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.399986 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.431815 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f1a88d0-7f56-47f4-874d-2c6023b235c1\") pod \"ovsdbserver-nb-1\" (UID: \"2e0f5920-9f22-4288-b09f-be3e565d88df\") " pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.443049 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.445069 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-982e641a-249c-4eed-bf2c-bea1e6987f2a\") pod \"ovsdbserver-nb-2\" (UID: \"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4\") " pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.493148 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:15 crc kubenswrapper[4979]: I1204 13:10:15.729364 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:16 crc kubenswrapper[4979]: I1204 13:10:16.036880 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-1"] Dec 04 13:10:16 crc kubenswrapper[4979]: I1204 13:10:16.119922 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 04 13:10:16 crc kubenswrapper[4979]: W1204 13:10:16.132010 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62c0f621_2918_4193_9420_780b3ecbd5e5.slice/crio-4333e70dd23e9aa4d214b1d2f851ffef991698ef09270d7f1f9414451870cf51 WatchSource:0}: Error finding container 4333e70dd23e9aa4d214b1d2f851ffef991698ef09270d7f1f9414451870cf51: Status 404 returned error can't find the container with id 4333e70dd23e9aa4d214b1d2f851ffef991698ef09270d7f1f9414451870cf51 Dec 04 13:10:16 crc kubenswrapper[4979]: I1204 13:10:16.451193 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-2"] Dec 04 13:10:16 crc kubenswrapper[4979]: I1204 13:10:16.652766 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-2"] Dec 04 13:10:16 crc kubenswrapper[4979]: W1204 13:10:16.659527 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd92753a_87a9_47ca_ae6b_add04cbf46d3.slice/crio-d3049660faa06f04190aad2938f485f809a2fbd09be3d1246d1a48361131d648 WatchSource:0}: Error finding container d3049660faa06f04190aad2938f485f809a2fbd09be3d1246d1a48361131d648: Status 404 returned error can't find the container with id d3049660faa06f04190aad2938f485f809a2fbd09be3d1246d1a48361131d648 Dec 04 13:10:16 crc kubenswrapper[4979]: I1204 13:10:16.961403 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-1"] Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.026790 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4","Type":"ContainerStarted","Data":"f8219f9e37a81cf44d6ad427ca4ae2f4c0873706bdc18bcf7af89544f77e9108"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.026832 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4","Type":"ContainerStarted","Data":"03dcf8e9eda61a5385f1301dea386ce5333d2075c80c0067bcbd4a8317bafe2a"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.026847 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-2" event={"ID":"d3ac5ba2-798d-4101-ad63-14cf9cfd47b4","Type":"ContainerStarted","Data":"2671891957f114bd2651a10bfe0236ae43149d1bab1bbc215ff049d905323609"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.028940 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"4d9841cd-dc04-4f50-8436-3d020df5f2f2","Type":"ContainerStarted","Data":"a3cf80da021c89350e6a348d8d63db5bce67761b3a02106e798cabdb6d76e5f0"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.028980 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"4d9841cd-dc04-4f50-8436-3d020df5f2f2","Type":"ContainerStarted","Data":"4ee3482554fc5bb24c69a3985ca92f7112bc6fc453e4a2f754522966a092a5de"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.028991 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-1" event={"ID":"4d9841cd-dc04-4f50-8436-3d020df5f2f2","Type":"ContainerStarted","Data":"bd27517577bb363891c533a460bbaa61266fed81e44b4e582bdc341ba4705c3e"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.031680 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"dd92753a-87a9-47ca-ae6b-add04cbf46d3","Type":"ContainerStarted","Data":"028c5d72ff6bcf2811ab5f55be1aa768ed17e3ad0e9fc5e933013d298a081860"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.031732 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"dd92753a-87a9-47ca-ae6b-add04cbf46d3","Type":"ContainerStarted","Data":"ff02a646bfa585a0c607fd339a5919b629f23fa994826b820ceb79da6e772bd1"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.031744 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-2" event={"ID":"dd92753a-87a9-47ca-ae6b-add04cbf46d3","Type":"ContainerStarted","Data":"d3049660faa06f04190aad2938f485f809a2fbd09be3d1246d1a48361131d648"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.034266 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"62c0f621-2918-4193-9420-780b3ecbd5e5","Type":"ContainerStarted","Data":"f56f114c5512b6937a9160eae456e8b9e419b22ef4f757798d8780ec50a3beb0"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.034316 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"62c0f621-2918-4193-9420-780b3ecbd5e5","Type":"ContainerStarted","Data":"fb4cf3b16663e4928b31abec48196b081324ccc2fae8e330c5a0d0a933b5b9f5"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.034327 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"62c0f621-2918-4193-9420-780b3ecbd5e5","Type":"ContainerStarted","Data":"4333e70dd23e9aa4d214b1d2f851ffef991698ef09270d7f1f9414451870cf51"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.036654 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"2e0f5920-9f22-4288-b09f-be3e565d88df","Type":"ContainerStarted","Data":"e1f878318cbce57f550a1b7f1b9d4c663acb27698c60b09c13f78f37c4d60c20"} Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.054289 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-2" podStartSLOduration=3.054267114 podStartE2EDuration="3.054267114s" podCreationTimestamp="2025-12-04 13:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:17.050711459 +0000 UTC m=+5241.325007283" watchObservedRunningTime="2025-12-04 13:10:17.054267114 +0000 UTC m=+5241.328562918" Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.076118 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=3.076101844 podStartE2EDuration="3.076101844s" podCreationTimestamp="2025-12-04 13:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:17.072776366 +0000 UTC m=+5241.347072170" watchObservedRunningTime="2025-12-04 13:10:17.076101844 +0000 UTC m=+5241.350397648" Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.096250 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-2" podStartSLOduration=4.09623115 podStartE2EDuration="4.09623115s" podCreationTimestamp="2025-12-04 13:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:17.091728479 +0000 UTC m=+5241.366024293" watchObservedRunningTime="2025-12-04 13:10:17.09623115 +0000 UTC m=+5241.370526954" Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.116193 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-1" podStartSLOduration=4.11616625 podStartE2EDuration="4.11616625s" podCreationTimestamp="2025-12-04 13:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:17.107410057 +0000 UTC m=+5241.381705861" watchObservedRunningTime="2025-12-04 13:10:17.11616625 +0000 UTC m=+5241.390462054" Dec 04 13:10:17 crc kubenswrapper[4979]: I1204 13:10:17.286434 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 04 13:10:17 crc kubenswrapper[4979]: W1204 13:10:17.287471 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6daaa5ce_1809_4b88_bec5_90204cbc1491.slice/crio-1051653aff92deeee320498b54263b3ca648b21b6d794bff3c7983fb655f47e1 WatchSource:0}: Error finding container 1051653aff92deeee320498b54263b3ca648b21b6d794bff3c7983fb655f47e1: Status 404 returned error can't find the container with id 1051653aff92deeee320498b54263b3ca648b21b6d794bff3c7983fb655f47e1 Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.049383 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"2e0f5920-9f22-4288-b09f-be3e565d88df","Type":"ContainerStarted","Data":"804db3d7292b75627e58e6e604964cf58c33ada9c382d2239a4a46d5a7ce484d"} Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.049768 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-1" event={"ID":"2e0f5920-9f22-4288-b09f-be3e565d88df","Type":"ContainerStarted","Data":"0220d4647ce26e29c755281c07ec3d52c5c5d9d149d7792b9f66ec580d4f7dcc"} Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.052209 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6daaa5ce-1809-4b88-bec5-90204cbc1491","Type":"ContainerStarted","Data":"7aee4ed33874298308f425f9b4ccfd2ecbf7c3cf326c214df3e94b48b53562c2"} Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.052268 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6daaa5ce-1809-4b88-bec5-90204cbc1491","Type":"ContainerStarted","Data":"33a47f3c8cab5f3b57f7ce20aad8d746a81cc4835643c4a9a1f029e950e5e6e2"} Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.052283 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"6daaa5ce-1809-4b88-bec5-90204cbc1491","Type":"ContainerStarted","Data":"1051653aff92deeee320498b54263b3ca648b21b6d794bff3c7983fb655f47e1"} Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.071076 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-1" podStartSLOduration=4.071055964 podStartE2EDuration="4.071055964s" podCreationTimestamp="2025-12-04 13:10:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:18.066673008 +0000 UTC m=+5242.340968832" watchObservedRunningTime="2025-12-04 13:10:18.071055964 +0000 UTC m=+5242.345351768" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.089088 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=5.089069333 podStartE2EDuration="5.089069333s" podCreationTimestamp="2025-12-04 13:10:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:18.086810583 +0000 UTC m=+5242.361106387" watchObservedRunningTime="2025-12-04 13:10:18.089069333 +0000 UTC m=+5242.363365137" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.240982 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.260662 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.400509 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.443410 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.493815 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:18 crc kubenswrapper[4979]: I1204 13:10:18.730377 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:20 crc kubenswrapper[4979]: I1204 13:10:20.240559 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:20 crc kubenswrapper[4979]: I1204 13:10:20.261213 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:20 crc kubenswrapper[4979]: I1204 13:10:20.400765 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:20 crc kubenswrapper[4979]: I1204 13:10:20.443994 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:20 crc kubenswrapper[4979]: I1204 13:10:20.494736 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:20 crc kubenswrapper[4979]: I1204 13:10:20.730327 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.283497 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.307236 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.355452 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-2" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.370345 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-1" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.456412 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.509660 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.578928 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfd7c6857-5b5bm"] Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.580338 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.584489 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.584503 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.609445 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.612861 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6rv8q\" (UniqueName: \"kubernetes.io/projected/5479c689-2415-4137-a37c-0e22e3ddb389-kube-api-access-6rv8q\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.612953 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-config\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.613022 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-dns-svc\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.613096 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-ovsdbserver-sb\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.615016 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfd7c6857-5b5bm"] Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.714681 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-ovsdbserver-sb\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.714755 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6rv8q\" (UniqueName: \"kubernetes.io/projected/5479c689-2415-4137-a37c-0e22e3ddb389-kube-api-access-6rv8q\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.714832 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-config\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.714888 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-dns-svc\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.715840 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-ovsdbserver-sb\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.716057 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-config\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.716149 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-dns-svc\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.738068 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6rv8q\" (UniqueName: \"kubernetes.io/projected/5479c689-2415-4137-a37c-0e22e3ddb389-kube-api-access-6rv8q\") pod \"dnsmasq-dns-7bfd7c6857-5b5bm\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.783623 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.829636 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-2" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.908362 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:21 crc kubenswrapper[4979]: I1204 13:10:21.955548 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfd7c6857-5b5bm"] Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.004218 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bff787dc9-ff85x"] Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.008079 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.019439 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-config\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.020031 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-nb\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.020085 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqd6q\" (UniqueName: \"kubernetes.io/projected/e9d44c50-a0aa-4aed-a2d0-ec957330b617-kube-api-access-qqd6q\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.020180 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-sb\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.020263 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-dns-svc\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.021993 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.023032 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bff787dc9-ff85x"] Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.122659 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-nb\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.122702 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqd6q\" (UniqueName: \"kubernetes.io/projected/e9d44c50-a0aa-4aed-a2d0-ec957330b617-kube-api-access-qqd6q\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.122793 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-sb\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.122843 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-dns-svc\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.122894 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-config\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.124889 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-config\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.125267 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-nb\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.125498 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-sb\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.126481 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-dns-svc\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.143183 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqd6q\" (UniqueName: \"kubernetes.io/projected/e9d44c50-a0aa-4aed-a2d0-ec957330b617-kube-api-access-qqd6q\") pod \"dnsmasq-dns-5bff787dc9-ff85x\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.158788 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.182473 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-1" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.348002 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.479974 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfd7c6857-5b5bm"] Dec 04 13:10:22 crc kubenswrapper[4979]: I1204 13:10:22.783817 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bff787dc9-ff85x"] Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.098581 4979 generic.go:334] "Generic (PLEG): container finished" podID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerID="a1b27d160ea952de7f5474594e3538fb6f02ad6224f6fd9996061e2e25e596ac" exitCode=0 Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.098645 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" event={"ID":"e9d44c50-a0aa-4aed-a2d0-ec957330b617","Type":"ContainerDied","Data":"a1b27d160ea952de7f5474594e3538fb6f02ad6224f6fd9996061e2e25e596ac"} Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.098704 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" event={"ID":"e9d44c50-a0aa-4aed-a2d0-ec957330b617","Type":"ContainerStarted","Data":"5185e6eca431efd3adc5a318b9f545645f6ea6031e923a0c43b4e667a456fc35"} Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.101072 4979 generic.go:334] "Generic (PLEG): container finished" podID="5479c689-2415-4137-a37c-0e22e3ddb389" containerID="f82ea6f42ca707f4f33b1c03b0f379cdd9dba3d91e6fee20093306914afcc169" exitCode=0 Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.101127 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" event={"ID":"5479c689-2415-4137-a37c-0e22e3ddb389","Type":"ContainerDied","Data":"f82ea6f42ca707f4f33b1c03b0f379cdd9dba3d91e6fee20093306914afcc169"} Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.101163 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" event={"ID":"5479c689-2415-4137-a37c-0e22e3ddb389","Type":"ContainerStarted","Data":"c1ad308eba52c7b231014364a987a1886ec1ac6bb8f0c90b1232cf1f7668ae21"} Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.370219 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.447689 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-dns-svc\") pod \"5479c689-2415-4137-a37c-0e22e3ddb389\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.447777 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-config\") pod \"5479c689-2415-4137-a37c-0e22e3ddb389\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.447820 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-ovsdbserver-sb\") pod \"5479c689-2415-4137-a37c-0e22e3ddb389\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.448529 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rv8q\" (UniqueName: \"kubernetes.io/projected/5479c689-2415-4137-a37c-0e22e3ddb389-kube-api-access-6rv8q\") pod \"5479c689-2415-4137-a37c-0e22e3ddb389\" (UID: \"5479c689-2415-4137-a37c-0e22e3ddb389\") " Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.451410 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5479c689-2415-4137-a37c-0e22e3ddb389-kube-api-access-6rv8q" (OuterVolumeSpecName: "kube-api-access-6rv8q") pod "5479c689-2415-4137-a37c-0e22e3ddb389" (UID: "5479c689-2415-4137-a37c-0e22e3ddb389"). InnerVolumeSpecName "kube-api-access-6rv8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.466105 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5479c689-2415-4137-a37c-0e22e3ddb389" (UID: "5479c689-2415-4137-a37c-0e22e3ddb389"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.466710 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5479c689-2415-4137-a37c-0e22e3ddb389" (UID: "5479c689-2415-4137-a37c-0e22e3ddb389"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.466754 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-config" (OuterVolumeSpecName: "config") pod "5479c689-2415-4137-a37c-0e22e3ddb389" (UID: "5479c689-2415-4137-a37c-0e22e3ddb389"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.551025 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6rv8q\" (UniqueName: \"kubernetes.io/projected/5479c689-2415-4137-a37c-0e22e3ddb389-kube-api-access-6rv8q\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.551067 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.551079 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:23 crc kubenswrapper[4979]: I1204 13:10:23.551089 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5479c689-2415-4137-a37c-0e22e3ddb389-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.110637 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" event={"ID":"5479c689-2415-4137-a37c-0e22e3ddb389","Type":"ContainerDied","Data":"c1ad308eba52c7b231014364a987a1886ec1ac6bb8f0c90b1232cf1f7668ae21"} Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.110673 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfd7c6857-5b5bm" Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.110719 4979 scope.go:117] "RemoveContainer" containerID="f82ea6f42ca707f4f33b1c03b0f379cdd9dba3d91e6fee20093306914afcc169" Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.115880 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" event={"ID":"e9d44c50-a0aa-4aed-a2d0-ec957330b617","Type":"ContainerStarted","Data":"6cfa180fcc4b2f4c636eba9b60cb0273ebee86ce6190eea95f74161b0a9fde06"} Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.116071 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.141501 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" podStartSLOduration=3.141480749 podStartE2EDuration="3.141480749s" podCreationTimestamp="2025-12-04 13:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:24.140690168 +0000 UTC m=+5248.414985972" watchObservedRunningTime="2025-12-04 13:10:24.141480749 +0000 UTC m=+5248.415776553" Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.197683 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfd7c6857-5b5bm"] Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.200123 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:10:24 crc kubenswrapper[4979]: E1204 13:10:24.200396 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:10:24 crc kubenswrapper[4979]: I1204 13:10:24.213006 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfd7c6857-5b5bm"] Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.260921 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-copy-data"] Dec 04 13:10:25 crc kubenswrapper[4979]: E1204 13:10:25.262126 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5479c689-2415-4137-a37c-0e22e3ddb389" containerName="init" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.262151 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5479c689-2415-4137-a37c-0e22e3ddb389" containerName="init" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.262342 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5479c689-2415-4137-a37c-0e22e3ddb389" containerName="init" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.263023 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.266865 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovn-data-cert" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.272353 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.380877 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.380935 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0856a4c9-e000-4d68-b30e-54d80b3afd79-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.381003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45wzf\" (UniqueName: \"kubernetes.io/projected/0856a4c9-e000-4d68-b30e-54d80b3afd79-kube-api-access-45wzf\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.482232 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.482518 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0856a4c9-e000-4d68-b30e-54d80b3afd79-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.482579 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45wzf\" (UniqueName: \"kubernetes.io/projected/0856a4c9-e000-4d68-b30e-54d80b3afd79-kube-api-access-45wzf\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.485246 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.485292 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/43c056687643562106f653ee717f2076ed69f7cb4cc518b1316fe3f3af9ef7ac/globalmount\"" pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.489406 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0856a4c9-e000-4d68-b30e-54d80b3afd79-ovn-data-cert\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.502727 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45wzf\" (UniqueName: \"kubernetes.io/projected/0856a4c9-e000-4d68-b30e-54d80b3afd79-kube-api-access-45wzf\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.516714 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") pod \"ovn-copy-data\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " pod="openstack/ovn-copy-data" Dec 04 13:10:25 crc kubenswrapper[4979]: I1204 13:10:25.586830 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 13:10:26 crc kubenswrapper[4979]: I1204 13:10:26.088961 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 13:10:26 crc kubenswrapper[4979]: I1204 13:10:26.134080 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0856a4c9-e000-4d68-b30e-54d80b3afd79","Type":"ContainerStarted","Data":"3a8d0b0446e6b1590ec3966f962209ce05177b721c7a0956b41268cc1dfd63a2"} Dec 04 13:10:26 crc kubenswrapper[4979]: I1204 13:10:26.209281 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5479c689-2415-4137-a37c-0e22e3ddb389" path="/var/lib/kubelet/pods/5479c689-2415-4137-a37c-0e22e3ddb389/volumes" Dec 04 13:10:27 crc kubenswrapper[4979]: I1204 13:10:27.143610 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0856a4c9-e000-4d68-b30e-54d80b3afd79","Type":"ContainerStarted","Data":"616e6cf4173e1da9dee11314ad17cf9d23935e33fb285e5e470c4ca84cb5e3e8"} Dec 04 13:10:27 crc kubenswrapper[4979]: I1204 13:10:27.165954 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-copy-data" podStartSLOduration=3.165914999 podStartE2EDuration="3.165914999s" podCreationTimestamp="2025-12-04 13:10:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:27.156350715 +0000 UTC m=+5251.430646519" watchObservedRunningTime="2025-12-04 13:10:27.165914999 +0000 UTC m=+5251.440210813" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.349525 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.432457 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-w5b8h"] Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.433040 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerName="dnsmasq-dns" containerID="cri-o://dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d" gracePeriod=10 Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.628972 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.641167 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.643940 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-92xnj" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.644138 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.644318 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.646655 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.708140 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ab214c3-6f92-443f-81e7-c48d306eba45-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.708221 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ab214c3-6f92-443f-81e7-c48d306eba45-config\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.708259 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ab214c3-6f92-443f-81e7-c48d306eba45-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.708390 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh29t\" (UniqueName: \"kubernetes.io/projected/0ab214c3-6f92-443f-81e7-c48d306eba45-kube-api-access-xh29t\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.708418 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ab214c3-6f92-443f-81e7-c48d306eba45-scripts\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.813687 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ab214c3-6f92-443f-81e7-c48d306eba45-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.813753 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ab214c3-6f92-443f-81e7-c48d306eba45-config\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.813777 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ab214c3-6f92-443f-81e7-c48d306eba45-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.813802 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh29t\" (UniqueName: \"kubernetes.io/projected/0ab214c3-6f92-443f-81e7-c48d306eba45-kube-api-access-xh29t\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.813820 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ab214c3-6f92-443f-81e7-c48d306eba45-scripts\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.814827 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0ab214c3-6f92-443f-81e7-c48d306eba45-scripts\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.814980 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ab214c3-6f92-443f-81e7-c48d306eba45-config\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.815326 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/0ab214c3-6f92-443f-81e7-c48d306eba45-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.825544 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ab214c3-6f92-443f-81e7-c48d306eba45-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.835413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh29t\" (UniqueName: \"kubernetes.io/projected/0ab214c3-6f92-443f-81e7-c48d306eba45-kube-api-access-xh29t\") pod \"ovn-northd-0\" (UID: \"0ab214c3-6f92-443f-81e7-c48d306eba45\") " pod="openstack/ovn-northd-0" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.937954 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:10:32 crc kubenswrapper[4979]: I1204 13:10:32.966600 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.018371 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-config\") pod \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.018516 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tx9x\" (UniqueName: \"kubernetes.io/projected/6c36cad8-31ac-477b-98f4-be4b4eeac17d-kube-api-access-9tx9x\") pod \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.018546 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-dns-svc\") pod \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\" (UID: \"6c36cad8-31ac-477b-98f4-be4b4eeac17d\") " Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.027166 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c36cad8-31ac-477b-98f4-be4b4eeac17d-kube-api-access-9tx9x" (OuterVolumeSpecName: "kube-api-access-9tx9x") pod "6c36cad8-31ac-477b-98f4-be4b4eeac17d" (UID: "6c36cad8-31ac-477b-98f4-be4b4eeac17d"). InnerVolumeSpecName "kube-api-access-9tx9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.063998 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-config" (OuterVolumeSpecName: "config") pod "6c36cad8-31ac-477b-98f4-be4b4eeac17d" (UID: "6c36cad8-31ac-477b-98f4-be4b4eeac17d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.090704 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6c36cad8-31ac-477b-98f4-be4b4eeac17d" (UID: "6c36cad8-31ac-477b-98f4-be4b4eeac17d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.122246 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tx9x\" (UniqueName: \"kubernetes.io/projected/6c36cad8-31ac-477b-98f4-be4b4eeac17d-kube-api-access-9tx9x\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.122291 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.122324 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c36cad8-31ac-477b-98f4-be4b4eeac17d-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.227769 4979 generic.go:334] "Generic (PLEG): container finished" podID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerID="dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d" exitCode=0 Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.227821 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" event={"ID":"6c36cad8-31ac-477b-98f4-be4b4eeac17d","Type":"ContainerDied","Data":"dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d"} Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.227849 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" event={"ID":"6c36cad8-31ac-477b-98f4-be4b4eeac17d","Type":"ContainerDied","Data":"e11f6121e80629a53531bd50a35efc7593d7d5ec6810c6b4eea0ad6a872ec925"} Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.227866 4979 scope.go:117] "RemoveContainer" containerID="dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.228032 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b7946d7b9-w5b8h" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.302926 4979 scope.go:117] "RemoveContainer" containerID="62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.304766 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-w5b8h"] Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.311559 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b7946d7b9-w5b8h"] Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.325273 4979 scope.go:117] "RemoveContainer" containerID="dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d" Dec 04 13:10:33 crc kubenswrapper[4979]: E1204 13:10:33.326115 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d\": container with ID starting with dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d not found: ID does not exist" containerID="dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.326272 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d"} err="failed to get container status \"dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d\": rpc error: code = NotFound desc = could not find container \"dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d\": container with ID starting with dbfdc27d3cf5e7558c0412a84098cdd34537a2aafc1964a9ab14cf69871bab7d not found: ID does not exist" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.326401 4979 scope.go:117] "RemoveContainer" containerID="62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4" Dec 04 13:10:33 crc kubenswrapper[4979]: E1204 13:10:33.326866 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4\": container with ID starting with 62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4 not found: ID does not exist" containerID="62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.326907 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4"} err="failed to get container status \"62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4\": rpc error: code = NotFound desc = could not find container \"62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4\": container with ID starting with 62e102d05f1e0b5418ba9bd00327fafed7880ceacc058a0b5632c72910cb33a4 not found: ID does not exist" Dec 04 13:10:33 crc kubenswrapper[4979]: I1204 13:10:33.493821 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 04 13:10:33 crc kubenswrapper[4979]: W1204 13:10:33.499128 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ab214c3_6f92_443f_81e7_c48d306eba45.slice/crio-f42dc3c4fb461ea5741f7df48c60af235c7ee18a2b0cf4b1848dd85068e315a2 WatchSource:0}: Error finding container f42dc3c4fb461ea5741f7df48c60af235c7ee18a2b0cf4b1848dd85068e315a2: Status 404 returned error can't find the container with id f42dc3c4fb461ea5741f7df48c60af235c7ee18a2b0cf4b1848dd85068e315a2 Dec 04 13:10:34 crc kubenswrapper[4979]: I1204 13:10:34.210833 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" path="/var/lib/kubelet/pods/6c36cad8-31ac-477b-98f4-be4b4eeac17d/volumes" Dec 04 13:10:34 crc kubenswrapper[4979]: I1204 13:10:34.240729 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0ab214c3-6f92-443f-81e7-c48d306eba45","Type":"ContainerStarted","Data":"bb0e665c6f62bc6226c0b24e4d341019990031970d3ddb68be63084a3df10566"} Dec 04 13:10:34 crc kubenswrapper[4979]: I1204 13:10:34.240785 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0ab214c3-6f92-443f-81e7-c48d306eba45","Type":"ContainerStarted","Data":"1d0f1ba3ca8e6df408e18494ce9007e3ec175932e9063d3e1ec97b0c803765c2"} Dec 04 13:10:34 crc kubenswrapper[4979]: I1204 13:10:34.240796 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"0ab214c3-6f92-443f-81e7-c48d306eba45","Type":"ContainerStarted","Data":"f42dc3c4fb461ea5741f7df48c60af235c7ee18a2b0cf4b1848dd85068e315a2"} Dec 04 13:10:34 crc kubenswrapper[4979]: I1204 13:10:34.240884 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 04 13:10:34 crc kubenswrapper[4979]: I1204 13:10:34.266595 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.266573531 podStartE2EDuration="2.266573531s" podCreationTimestamp="2025-12-04 13:10:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:34.262841191 +0000 UTC m=+5258.537136995" watchObservedRunningTime="2025-12-04 13:10:34.266573531 +0000 UTC m=+5258.540869335" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.199829 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:10:37 crc kubenswrapper[4979]: E1204 13:10:37.200403 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.929849 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-xckk8"] Dec 04 13:10:37 crc kubenswrapper[4979]: E1204 13:10:37.930559 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerName="dnsmasq-dns" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.930651 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerName="dnsmasq-dns" Dec 04 13:10:37 crc kubenswrapper[4979]: E1204 13:10:37.930728 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerName="init" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.930832 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerName="init" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.931110 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c36cad8-31ac-477b-98f4-be4b4eeac17d" containerName="dnsmasq-dns" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.931916 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:37 crc kubenswrapper[4979]: I1204 13:10:37.942619 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xckk8"] Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.020980 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-0762-account-create-update-f26mh"] Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.022344 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.033710 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.034116 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0762-account-create-update-f26mh"] Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.099334 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bndkz\" (UniqueName: \"kubernetes.io/projected/e57348b9-9bff-4b87-8075-b8bf9e459d02-kube-api-access-bndkz\") pod \"keystone-db-create-xckk8\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.099430 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57348b9-9bff-4b87-8075-b8bf9e459d02-operator-scripts\") pod \"keystone-db-create-xckk8\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.200224 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bndkz\" (UniqueName: \"kubernetes.io/projected/e57348b9-9bff-4b87-8075-b8bf9e459d02-kube-api-access-bndkz\") pod \"keystone-db-create-xckk8\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.200335 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/917d86c8-d0fd-4769-aa6f-f1bc490c1406-operator-scripts\") pod \"keystone-0762-account-create-update-f26mh\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.200410 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9cst\" (UniqueName: \"kubernetes.io/projected/917d86c8-d0fd-4769-aa6f-f1bc490c1406-kube-api-access-s9cst\") pod \"keystone-0762-account-create-update-f26mh\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.200449 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57348b9-9bff-4b87-8075-b8bf9e459d02-operator-scripts\") pod \"keystone-db-create-xckk8\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.201198 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57348b9-9bff-4b87-8075-b8bf9e459d02-operator-scripts\") pod \"keystone-db-create-xckk8\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.224661 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bndkz\" (UniqueName: \"kubernetes.io/projected/e57348b9-9bff-4b87-8075-b8bf9e459d02-kube-api-access-bndkz\") pod \"keystone-db-create-xckk8\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.254515 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.302018 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/917d86c8-d0fd-4769-aa6f-f1bc490c1406-operator-scripts\") pod \"keystone-0762-account-create-update-f26mh\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.302125 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9cst\" (UniqueName: \"kubernetes.io/projected/917d86c8-d0fd-4769-aa6f-f1bc490c1406-kube-api-access-s9cst\") pod \"keystone-0762-account-create-update-f26mh\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.304908 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/917d86c8-d0fd-4769-aa6f-f1bc490c1406-operator-scripts\") pod \"keystone-0762-account-create-update-f26mh\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.325412 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9cst\" (UniqueName: \"kubernetes.io/projected/917d86c8-d0fd-4769-aa6f-f1bc490c1406-kube-api-access-s9cst\") pod \"keystone-0762-account-create-update-f26mh\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.350663 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.832549 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-xckk8"] Dec 04 13:10:38 crc kubenswrapper[4979]: W1204 13:10:38.833935 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode57348b9_9bff_4b87_8075_b8bf9e459d02.slice/crio-458943d9766e53ccc8b0b01f17099acdf017e14348d506a644540936d39feff8 WatchSource:0}: Error finding container 458943d9766e53ccc8b0b01f17099acdf017e14348d506a644540936d39feff8: Status 404 returned error can't find the container with id 458943d9766e53ccc8b0b01f17099acdf017e14348d506a644540936d39feff8 Dec 04 13:10:38 crc kubenswrapper[4979]: I1204 13:10:38.942086 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0762-account-create-update-f26mh"] Dec 04 13:10:38 crc kubenswrapper[4979]: W1204 13:10:38.946411 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod917d86c8_d0fd_4769_aa6f_f1bc490c1406.slice/crio-009bb8bb9f4c458bea58293087312987ddbcb118b7a9a5ac47ccf40060d6c275 WatchSource:0}: Error finding container 009bb8bb9f4c458bea58293087312987ddbcb118b7a9a5ac47ccf40060d6c275: Status 404 returned error can't find the container with id 009bb8bb9f4c458bea58293087312987ddbcb118b7a9a5ac47ccf40060d6c275 Dec 04 13:10:39 crc kubenswrapper[4979]: I1204 13:10:39.279491 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xckk8" event={"ID":"e57348b9-9bff-4b87-8075-b8bf9e459d02","Type":"ContainerDied","Data":"a9d1bdb9a9b4e6fb7af3f9d4801cc3463ca8134a3a4ccb3089d5cd0579d83cb9"} Dec 04 13:10:39 crc kubenswrapper[4979]: I1204 13:10:39.279359 4979 generic.go:334] "Generic (PLEG): container finished" podID="e57348b9-9bff-4b87-8075-b8bf9e459d02" containerID="a9d1bdb9a9b4e6fb7af3f9d4801cc3463ca8134a3a4ccb3089d5cd0579d83cb9" exitCode=0 Dec 04 13:10:39 crc kubenswrapper[4979]: I1204 13:10:39.280017 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xckk8" event={"ID":"e57348b9-9bff-4b87-8075-b8bf9e459d02","Type":"ContainerStarted","Data":"458943d9766e53ccc8b0b01f17099acdf017e14348d506a644540936d39feff8"} Dec 04 13:10:39 crc kubenswrapper[4979]: I1204 13:10:39.281983 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0762-account-create-update-f26mh" event={"ID":"917d86c8-d0fd-4769-aa6f-f1bc490c1406","Type":"ContainerStarted","Data":"b554a368d45951b3c430a573c940dd834684b37e97d23fc863303c50f2b7e2da"} Dec 04 13:10:39 crc kubenswrapper[4979]: I1204 13:10:39.282048 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0762-account-create-update-f26mh" event={"ID":"917d86c8-d0fd-4769-aa6f-f1bc490c1406","Type":"ContainerStarted","Data":"009bb8bb9f4c458bea58293087312987ddbcb118b7a9a5ac47ccf40060d6c275"} Dec 04 13:10:39 crc kubenswrapper[4979]: I1204 13:10:39.320376 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-0762-account-create-update-f26mh" podStartSLOduration=2.320351218 podStartE2EDuration="2.320351218s" podCreationTimestamp="2025-12-04 13:10:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:39.314183935 +0000 UTC m=+5263.588479759" watchObservedRunningTime="2025-12-04 13:10:39.320351218 +0000 UTC m=+5263.594647022" Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.290188 4979 generic.go:334] "Generic (PLEG): container finished" podID="917d86c8-d0fd-4769-aa6f-f1bc490c1406" containerID="b554a368d45951b3c430a573c940dd834684b37e97d23fc863303c50f2b7e2da" exitCode=0 Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.290230 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0762-account-create-update-f26mh" event={"ID":"917d86c8-d0fd-4769-aa6f-f1bc490c1406","Type":"ContainerDied","Data":"b554a368d45951b3c430a573c940dd834684b37e97d23fc863303c50f2b7e2da"} Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.624133 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.747434 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bndkz\" (UniqueName: \"kubernetes.io/projected/e57348b9-9bff-4b87-8075-b8bf9e459d02-kube-api-access-bndkz\") pod \"e57348b9-9bff-4b87-8075-b8bf9e459d02\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.747538 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57348b9-9bff-4b87-8075-b8bf9e459d02-operator-scripts\") pod \"e57348b9-9bff-4b87-8075-b8bf9e459d02\" (UID: \"e57348b9-9bff-4b87-8075-b8bf9e459d02\") " Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.748347 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e57348b9-9bff-4b87-8075-b8bf9e459d02-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e57348b9-9bff-4b87-8075-b8bf9e459d02" (UID: "e57348b9-9bff-4b87-8075-b8bf9e459d02"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.754092 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57348b9-9bff-4b87-8075-b8bf9e459d02-kube-api-access-bndkz" (OuterVolumeSpecName: "kube-api-access-bndkz") pod "e57348b9-9bff-4b87-8075-b8bf9e459d02" (UID: "e57348b9-9bff-4b87-8075-b8bf9e459d02"). InnerVolumeSpecName "kube-api-access-bndkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.849588 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bndkz\" (UniqueName: \"kubernetes.io/projected/e57348b9-9bff-4b87-8075-b8bf9e459d02-kube-api-access-bndkz\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:40 crc kubenswrapper[4979]: I1204 13:10:40.849635 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e57348b9-9bff-4b87-8075-b8bf9e459d02-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:41 crc kubenswrapper[4979]: I1204 13:10:41.300340 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-xckk8" event={"ID":"e57348b9-9bff-4b87-8075-b8bf9e459d02","Type":"ContainerDied","Data":"458943d9766e53ccc8b0b01f17099acdf017e14348d506a644540936d39feff8"} Dec 04 13:10:41 crc kubenswrapper[4979]: I1204 13:10:41.300703 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="458943d9766e53ccc8b0b01f17099acdf017e14348d506a644540936d39feff8" Dec 04 13:10:41 crc kubenswrapper[4979]: I1204 13:10:41.300409 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-xckk8" Dec 04 13:10:41 crc kubenswrapper[4979]: I1204 13:10:41.908934 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.073868 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/917d86c8-d0fd-4769-aa6f-f1bc490c1406-operator-scripts\") pod \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.074090 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9cst\" (UniqueName: \"kubernetes.io/projected/917d86c8-d0fd-4769-aa6f-f1bc490c1406-kube-api-access-s9cst\") pod \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\" (UID: \"917d86c8-d0fd-4769-aa6f-f1bc490c1406\") " Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.074493 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/917d86c8-d0fd-4769-aa6f-f1bc490c1406-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "917d86c8-d0fd-4769-aa6f-f1bc490c1406" (UID: "917d86c8-d0fd-4769-aa6f-f1bc490c1406"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.078291 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/917d86c8-d0fd-4769-aa6f-f1bc490c1406-kube-api-access-s9cst" (OuterVolumeSpecName: "kube-api-access-s9cst") pod "917d86c8-d0fd-4769-aa6f-f1bc490c1406" (UID: "917d86c8-d0fd-4769-aa6f-f1bc490c1406"). InnerVolumeSpecName "kube-api-access-s9cst". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.175670 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9cst\" (UniqueName: \"kubernetes.io/projected/917d86c8-d0fd-4769-aa6f-f1bc490c1406-kube-api-access-s9cst\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.175705 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/917d86c8-d0fd-4769-aa6f-f1bc490c1406-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.310362 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0762-account-create-update-f26mh" event={"ID":"917d86c8-d0fd-4769-aa6f-f1bc490c1406","Type":"ContainerDied","Data":"009bb8bb9f4c458bea58293087312987ddbcb118b7a9a5ac47ccf40060d6c275"} Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.310642 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="009bb8bb9f4c458bea58293087312987ddbcb118b7a9a5ac47ccf40060d6c275" Dec 04 13:10:42 crc kubenswrapper[4979]: I1204 13:10:42.310432 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0762-account-create-update-f26mh" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.531036 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-pqvcx"] Dec 04 13:10:43 crc kubenswrapper[4979]: E1204 13:10:43.531407 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="917d86c8-d0fd-4769-aa6f-f1bc490c1406" containerName="mariadb-account-create-update" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.531428 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="917d86c8-d0fd-4769-aa6f-f1bc490c1406" containerName="mariadb-account-create-update" Dec 04 13:10:43 crc kubenswrapper[4979]: E1204 13:10:43.531467 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57348b9-9bff-4b87-8075-b8bf9e459d02" containerName="mariadb-database-create" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.531477 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57348b9-9bff-4b87-8075-b8bf9e459d02" containerName="mariadb-database-create" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.531670 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57348b9-9bff-4b87-8075-b8bf9e459d02" containerName="mariadb-database-create" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.531693 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="917d86c8-d0fd-4769-aa6f-f1bc490c1406" containerName="mariadb-account-create-update" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.532330 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.534870 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.534920 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.534997 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4lfhw" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.535071 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.548183 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pqvcx"] Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.695820 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-config-data\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.696219 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvj8s\" (UniqueName: \"kubernetes.io/projected/6753d199-6125-4b42-bfa8-cae894e9aa06-kube-api-access-kvj8s\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.696473 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-combined-ca-bundle\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.797867 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-config-data\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.797931 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvj8s\" (UniqueName: \"kubernetes.io/projected/6753d199-6125-4b42-bfa8-cae894e9aa06-kube-api-access-kvj8s\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.798012 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-combined-ca-bundle\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.803245 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-config-data\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.803487 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-combined-ca-bundle\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.816022 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvj8s\" (UniqueName: \"kubernetes.io/projected/6753d199-6125-4b42-bfa8-cae894e9aa06-kube-api-access-kvj8s\") pod \"keystone-db-sync-pqvcx\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:43 crc kubenswrapper[4979]: I1204 13:10:43.850341 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:44 crc kubenswrapper[4979]: I1204 13:10:44.301416 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-pqvcx"] Dec 04 13:10:44 crc kubenswrapper[4979]: I1204 13:10:44.325412 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pqvcx" event={"ID":"6753d199-6125-4b42-bfa8-cae894e9aa06","Type":"ContainerStarted","Data":"88f902c9c90cd64109f6c12bdc6fae443831a1ab063aae5ceac7dc5c658aa99c"} Dec 04 13:10:46 crc kubenswrapper[4979]: I1204 13:10:46.338891 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pqvcx" event={"ID":"6753d199-6125-4b42-bfa8-cae894e9aa06","Type":"ContainerStarted","Data":"0951ba9ea4751f2febbfaac4bae14ed3efd49fbbdbf2a38c901e3db830f44fcd"} Dec 04 13:10:46 crc kubenswrapper[4979]: I1204 13:10:46.360400 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-pqvcx" podStartSLOduration=3.360382039 podStartE2EDuration="3.360382039s" podCreationTimestamp="2025-12-04 13:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:46.354233085 +0000 UTC m=+5270.628528899" watchObservedRunningTime="2025-12-04 13:10:46.360382039 +0000 UTC m=+5270.634677843" Dec 04 13:10:48 crc kubenswrapper[4979]: I1204 13:10:48.025440 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 04 13:10:48 crc kubenswrapper[4979]: I1204 13:10:48.355547 4979 generic.go:334] "Generic (PLEG): container finished" podID="6753d199-6125-4b42-bfa8-cae894e9aa06" containerID="0951ba9ea4751f2febbfaac4bae14ed3efd49fbbdbf2a38c901e3db830f44fcd" exitCode=0 Dec 04 13:10:48 crc kubenswrapper[4979]: I1204 13:10:48.355595 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pqvcx" event={"ID":"6753d199-6125-4b42-bfa8-cae894e9aa06","Type":"ContainerDied","Data":"0951ba9ea4751f2febbfaac4bae14ed3efd49fbbdbf2a38c901e3db830f44fcd"} Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.199040 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:10:49 crc kubenswrapper[4979]: E1204 13:10:49.199361 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.674950 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.789366 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvj8s\" (UniqueName: \"kubernetes.io/projected/6753d199-6125-4b42-bfa8-cae894e9aa06-kube-api-access-kvj8s\") pod \"6753d199-6125-4b42-bfa8-cae894e9aa06\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.789602 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-config-data\") pod \"6753d199-6125-4b42-bfa8-cae894e9aa06\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.789637 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-combined-ca-bundle\") pod \"6753d199-6125-4b42-bfa8-cae894e9aa06\" (UID: \"6753d199-6125-4b42-bfa8-cae894e9aa06\") " Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.808523 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6753d199-6125-4b42-bfa8-cae894e9aa06-kube-api-access-kvj8s" (OuterVolumeSpecName: "kube-api-access-kvj8s") pod "6753d199-6125-4b42-bfa8-cae894e9aa06" (UID: "6753d199-6125-4b42-bfa8-cae894e9aa06"). InnerVolumeSpecName "kube-api-access-kvj8s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.841952 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6753d199-6125-4b42-bfa8-cae894e9aa06" (UID: "6753d199-6125-4b42-bfa8-cae894e9aa06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.877751 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-config-data" (OuterVolumeSpecName: "config-data") pod "6753d199-6125-4b42-bfa8-cae894e9aa06" (UID: "6753d199-6125-4b42-bfa8-cae894e9aa06"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.891335 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvj8s\" (UniqueName: \"kubernetes.io/projected/6753d199-6125-4b42-bfa8-cae894e9aa06-kube-api-access-kvj8s\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.891375 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:49 crc kubenswrapper[4979]: I1204 13:10:49.891388 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6753d199-6125-4b42-bfa8-cae894e9aa06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.373717 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-pqvcx" event={"ID":"6753d199-6125-4b42-bfa8-cae894e9aa06","Type":"ContainerDied","Data":"88f902c9c90cd64109f6c12bdc6fae443831a1ab063aae5ceac7dc5c658aa99c"} Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.373769 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88f902c9c90cd64109f6c12bdc6fae443831a1ab063aae5ceac7dc5c658aa99c" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.373869 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-pqvcx" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.627771 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-96578f89-dn7t8"] Dec 04 13:10:50 crc kubenswrapper[4979]: E1204 13:10:50.628123 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6753d199-6125-4b42-bfa8-cae894e9aa06" containerName="keystone-db-sync" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.628135 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6753d199-6125-4b42-bfa8-cae894e9aa06" containerName="keystone-db-sync" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.628336 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6753d199-6125-4b42-bfa8-cae894e9aa06" containerName="keystone-db-sync" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.629198 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.644263 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-96578f89-dn7t8"] Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.687253 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-84lkr"] Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.689088 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.698595 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.698938 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.699411 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.699490 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4lfhw" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.699560 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.702199 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-84lkr"] Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.812432 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-scripts\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.812528 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-dns-svc\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.812620 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-nb\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.812654 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9gvt\" (UniqueName: \"kubernetes.io/projected/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-kube-api-access-b9gvt\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.812685 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-credential-keys\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.812972 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-sb\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.813058 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-config-data\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.813137 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-fernet-keys\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.813184 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8m9s\" (UniqueName: \"kubernetes.io/projected/aa4487ca-6eb4-4484-be18-48320bd5f6f9-kube-api-access-q8m9s\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.813258 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-combined-ca-bundle\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.813371 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-config\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-combined-ca-bundle\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915105 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-config\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915152 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-scripts\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915176 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-dns-svc\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915216 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-nb\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915240 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9gvt\" (UniqueName: \"kubernetes.io/projected/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-kube-api-access-b9gvt\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915264 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-credential-keys\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915369 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-sb\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915400 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-config-data\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-fernet-keys\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.915464 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8m9s\" (UniqueName: \"kubernetes.io/projected/aa4487ca-6eb4-4484-be18-48320bd5f6f9-kube-api-access-q8m9s\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.917888 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-dns-svc\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.918612 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-nb\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.919897 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-config\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.920556 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-combined-ca-bundle\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.920566 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-sb\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.922141 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-fernet-keys\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.922392 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-scripts\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.922852 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-credential-keys\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.923636 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-config-data\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.933527 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8m9s\" (UniqueName: \"kubernetes.io/projected/aa4487ca-6eb4-4484-be18-48320bd5f6f9-kube-api-access-q8m9s\") pod \"keystone-bootstrap-84lkr\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.942826 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9gvt\" (UniqueName: \"kubernetes.io/projected/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-kube-api-access-b9gvt\") pod \"dnsmasq-dns-96578f89-dn7t8\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:50 crc kubenswrapper[4979]: I1204 13:10:50.952347 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:51 crc kubenswrapper[4979]: I1204 13:10:51.037328 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:51 crc kubenswrapper[4979]: I1204 13:10:51.444277 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-96578f89-dn7t8"] Dec 04 13:10:51 crc kubenswrapper[4979]: I1204 13:10:51.569194 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-84lkr"] Dec 04 13:10:52 crc kubenswrapper[4979]: I1204 13:10:52.390733 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-84lkr" event={"ID":"aa4487ca-6eb4-4484-be18-48320bd5f6f9","Type":"ContainerStarted","Data":"8af1e4510309f447a9af3a1f372710c1c495d92cb253ea23e18a59d5bb24a7f2"} Dec 04 13:10:52 crc kubenswrapper[4979]: I1204 13:10:52.391085 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-84lkr" event={"ID":"aa4487ca-6eb4-4484-be18-48320bd5f6f9","Type":"ContainerStarted","Data":"f030b5438446e25900b7ed5bc26809175455cf75d012a1b7a076b4ff8d0f3847"} Dec 04 13:10:52 crc kubenswrapper[4979]: I1204 13:10:52.393423 4979 generic.go:334] "Generic (PLEG): container finished" podID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerID="45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580" exitCode=0 Dec 04 13:10:52 crc kubenswrapper[4979]: I1204 13:10:52.393474 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96578f89-dn7t8" event={"ID":"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee","Type":"ContainerDied","Data":"45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580"} Dec 04 13:10:52 crc kubenswrapper[4979]: I1204 13:10:52.393501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96578f89-dn7t8" event={"ID":"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee","Type":"ContainerStarted","Data":"3f01fc88f0e647edac4410bdba29d0c1e1c8573aecc958fa3554b59c77e81184"} Dec 04 13:10:52 crc kubenswrapper[4979]: I1204 13:10:52.415104 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-84lkr" podStartSLOduration=2.415081645 podStartE2EDuration="2.415081645s" podCreationTimestamp="2025-12-04 13:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:52.413970495 +0000 UTC m=+5276.688266289" watchObservedRunningTime="2025-12-04 13:10:52.415081645 +0000 UTC m=+5276.689377449" Dec 04 13:10:53 crc kubenswrapper[4979]: I1204 13:10:53.404619 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96578f89-dn7t8" event={"ID":"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee","Type":"ContainerStarted","Data":"088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5"} Dec 04 13:10:53 crc kubenswrapper[4979]: I1204 13:10:53.433005 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-96578f89-dn7t8" podStartSLOduration=3.432982154 podStartE2EDuration="3.432982154s" podCreationTimestamp="2025-12-04 13:10:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:10:53.424351035 +0000 UTC m=+5277.698646859" watchObservedRunningTime="2025-12-04 13:10:53.432982154 +0000 UTC m=+5277.707277958" Dec 04 13:10:54 crc kubenswrapper[4979]: I1204 13:10:54.434844 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:10:56 crc kubenswrapper[4979]: I1204 13:10:56.454993 4979 generic.go:334] "Generic (PLEG): container finished" podID="aa4487ca-6eb4-4484-be18-48320bd5f6f9" containerID="8af1e4510309f447a9af3a1f372710c1c495d92cb253ea23e18a59d5bb24a7f2" exitCode=0 Dec 04 13:10:56 crc kubenswrapper[4979]: I1204 13:10:56.455038 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-84lkr" event={"ID":"aa4487ca-6eb4-4484-be18-48320bd5f6f9","Type":"ContainerDied","Data":"8af1e4510309f447a9af3a1f372710c1c495d92cb253ea23e18a59d5bb24a7f2"} Dec 04 13:10:57 crc kubenswrapper[4979]: I1204 13:10:57.865319 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.044821 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-config-data\") pod \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.044888 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8m9s\" (UniqueName: \"kubernetes.io/projected/aa4487ca-6eb4-4484-be18-48320bd5f6f9-kube-api-access-q8m9s\") pod \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.044936 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-credential-keys\") pod \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.044973 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-combined-ca-bundle\") pod \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.045014 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-fernet-keys\") pod \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.045113 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-scripts\") pod \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\" (UID: \"aa4487ca-6eb4-4484-be18-48320bd5f6f9\") " Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.051377 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "aa4487ca-6eb4-4484-be18-48320bd5f6f9" (UID: "aa4487ca-6eb4-4484-be18-48320bd5f6f9"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.051602 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa4487ca-6eb4-4484-be18-48320bd5f6f9-kube-api-access-q8m9s" (OuterVolumeSpecName: "kube-api-access-q8m9s") pod "aa4487ca-6eb4-4484-be18-48320bd5f6f9" (UID: "aa4487ca-6eb4-4484-be18-48320bd5f6f9"). InnerVolumeSpecName "kube-api-access-q8m9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.052010 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-scripts" (OuterVolumeSpecName: "scripts") pod "aa4487ca-6eb4-4484-be18-48320bd5f6f9" (UID: "aa4487ca-6eb4-4484-be18-48320bd5f6f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.052077 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "aa4487ca-6eb4-4484-be18-48320bd5f6f9" (UID: "aa4487ca-6eb4-4484-be18-48320bd5f6f9"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.075226 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aa4487ca-6eb4-4484-be18-48320bd5f6f9" (UID: "aa4487ca-6eb4-4484-be18-48320bd5f6f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.076002 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-config-data" (OuterVolumeSpecName: "config-data") pod "aa4487ca-6eb4-4484-be18-48320bd5f6f9" (UID: "aa4487ca-6eb4-4484-be18-48320bd5f6f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.146936 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.146994 4979 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.147007 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.147018 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.147031 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8m9s\" (UniqueName: \"kubernetes.io/projected/aa4487ca-6eb4-4484-be18-48320bd5f6f9-kube-api-access-q8m9s\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.147046 4979 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/aa4487ca-6eb4-4484-be18-48320bd5f6f9-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.473146 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-84lkr" event={"ID":"aa4487ca-6eb4-4484-be18-48320bd5f6f9","Type":"ContainerDied","Data":"f030b5438446e25900b7ed5bc26809175455cf75d012a1b7a076b4ff8d0f3847"} Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.473199 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f030b5438446e25900b7ed5bc26809175455cf75d012a1b7a076b4ff8d0f3847" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.473211 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-84lkr" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.590615 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-84lkr"] Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.599217 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-84lkr"] Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.660329 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-958bg"] Dec 04 13:10:58 crc kubenswrapper[4979]: E1204 13:10:58.660717 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa4487ca-6eb4-4484-be18-48320bd5f6f9" containerName="keystone-bootstrap" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.660747 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa4487ca-6eb4-4484-be18-48320bd5f6f9" containerName="keystone-bootstrap" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.660957 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa4487ca-6eb4-4484-be18-48320bd5f6f9" containerName="keystone-bootstrap" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.661582 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.664562 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.664567 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.664742 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.664795 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4lfhw" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.665172 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.670569 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-958bg"] Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.756627 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-combined-ca-bundle\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.756954 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-fernet-keys\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.757054 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-config-data\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.757150 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-credential-keys\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.757265 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zktrz\" (UniqueName: \"kubernetes.io/projected/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-kube-api-access-zktrz\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.757394 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-scripts\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.859473 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-credential-keys\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.859605 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zktrz\" (UniqueName: \"kubernetes.io/projected/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-kube-api-access-zktrz\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.859650 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-scripts\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.859707 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-combined-ca-bundle\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.859741 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-fernet-keys\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.859764 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-config-data\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.863756 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-credential-keys\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.864152 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-scripts\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.864484 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-config-data\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.864780 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-combined-ca-bundle\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.864822 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-fernet-keys\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.879508 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zktrz\" (UniqueName: \"kubernetes.io/projected/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-kube-api-access-zktrz\") pod \"keystone-bootstrap-958bg\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:58 crc kubenswrapper[4979]: I1204 13:10:58.980875 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-958bg" Dec 04 13:10:59 crc kubenswrapper[4979]: I1204 13:10:59.458659 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-958bg"] Dec 04 13:10:59 crc kubenswrapper[4979]: W1204 13:10:59.466167 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5fa6e5e6_5309_489a_b5f7_1b6f34afac11.slice/crio-d568dd45ae2211eb0c50f0ab694d676dc60f83fad2b125ab477d40a2a526c448 WatchSource:0}: Error finding container d568dd45ae2211eb0c50f0ab694d676dc60f83fad2b125ab477d40a2a526c448: Status 404 returned error can't find the container with id d568dd45ae2211eb0c50f0ab694d676dc60f83fad2b125ab477d40a2a526c448 Dec 04 13:10:59 crc kubenswrapper[4979]: I1204 13:10:59.483653 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-958bg" event={"ID":"5fa6e5e6-5309-489a-b5f7-1b6f34afac11","Type":"ContainerStarted","Data":"d568dd45ae2211eb0c50f0ab694d676dc60f83fad2b125ab477d40a2a526c448"} Dec 04 13:11:00 crc kubenswrapper[4979]: I1204 13:11:00.209253 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa4487ca-6eb4-4484-be18-48320bd5f6f9" path="/var/lib/kubelet/pods/aa4487ca-6eb4-4484-be18-48320bd5f6f9/volumes" Dec 04 13:11:00 crc kubenswrapper[4979]: I1204 13:11:00.496034 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-958bg" event={"ID":"5fa6e5e6-5309-489a-b5f7-1b6f34afac11","Type":"ContainerStarted","Data":"a4960334836f7e198fd176ea85315ee9f98088e691848d494e35542d764e7349"} Dec 04 13:11:00 crc kubenswrapper[4979]: I1204 13:11:00.522873 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-958bg" podStartSLOduration=2.52284862 podStartE2EDuration="2.52284862s" podCreationTimestamp="2025-12-04 13:10:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:11:00.518784061 +0000 UTC m=+5284.793079865" watchObservedRunningTime="2025-12-04 13:11:00.52284862 +0000 UTC m=+5284.797144424" Dec 04 13:11:00 crc kubenswrapper[4979]: I1204 13:11:00.954284 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.039050 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bff787dc9-ff85x"] Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.039339 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerName="dnsmasq-dns" containerID="cri-o://6cfa180fcc4b2f4c636eba9b60cb0273ebee86ce6190eea95f74161b0a9fde06" gracePeriod=10 Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.529523 4979 generic.go:334] "Generic (PLEG): container finished" podID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerID="6cfa180fcc4b2f4c636eba9b60cb0273ebee86ce6190eea95f74161b0a9fde06" exitCode=0 Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.529985 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" event={"ID":"e9d44c50-a0aa-4aed-a2d0-ec957330b617","Type":"ContainerDied","Data":"6cfa180fcc4b2f4c636eba9b60cb0273ebee86ce6190eea95f74161b0a9fde06"} Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.530054 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" event={"ID":"e9d44c50-a0aa-4aed-a2d0-ec957330b617","Type":"ContainerDied","Data":"5185e6eca431efd3adc5a318b9f545645f6ea6031e923a0c43b4e667a456fc35"} Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.530070 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5185e6eca431efd3adc5a318b9f545645f6ea6031e923a0c43b4e667a456fc35" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.570929 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.714351 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-dns-svc\") pod \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.714433 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-config\") pod \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.714574 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqd6q\" (UniqueName: \"kubernetes.io/projected/e9d44c50-a0aa-4aed-a2d0-ec957330b617-kube-api-access-qqd6q\") pod \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.714617 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-sb\") pod \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.714718 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-nb\") pod \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\" (UID: \"e9d44c50-a0aa-4aed-a2d0-ec957330b617\") " Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.742571 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9d44c50-a0aa-4aed-a2d0-ec957330b617-kube-api-access-qqd6q" (OuterVolumeSpecName: "kube-api-access-qqd6q") pod "e9d44c50-a0aa-4aed-a2d0-ec957330b617" (UID: "e9d44c50-a0aa-4aed-a2d0-ec957330b617"). InnerVolumeSpecName "kube-api-access-qqd6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.776226 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e9d44c50-a0aa-4aed-a2d0-ec957330b617" (UID: "e9d44c50-a0aa-4aed-a2d0-ec957330b617"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.792146 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-config" (OuterVolumeSpecName: "config") pod "e9d44c50-a0aa-4aed-a2d0-ec957330b617" (UID: "e9d44c50-a0aa-4aed-a2d0-ec957330b617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.815116 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e9d44c50-a0aa-4aed-a2d0-ec957330b617" (UID: "e9d44c50-a0aa-4aed-a2d0-ec957330b617"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.818074 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.818113 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqd6q\" (UniqueName: \"kubernetes.io/projected/e9d44c50-a0aa-4aed-a2d0-ec957330b617-kube-api-access-qqd6q\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.818128 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.818140 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.835975 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e9d44c50-a0aa-4aed-a2d0-ec957330b617" (UID: "e9d44c50-a0aa-4aed-a2d0-ec957330b617"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:11:01 crc kubenswrapper[4979]: I1204 13:11:01.919163 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e9d44c50-a0aa-4aed-a2d0-ec957330b617-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:02 crc kubenswrapper[4979]: I1204 13:11:02.536251 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bff787dc9-ff85x" Dec 04 13:11:02 crc kubenswrapper[4979]: I1204 13:11:02.563943 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bff787dc9-ff85x"] Dec 04 13:11:02 crc kubenswrapper[4979]: I1204 13:11:02.592461 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bff787dc9-ff85x"] Dec 04 13:11:03 crc kubenswrapper[4979]: I1204 13:11:03.199410 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:11:03 crc kubenswrapper[4979]: I1204 13:11:03.546406 4979 generic.go:334] "Generic (PLEG): container finished" podID="5fa6e5e6-5309-489a-b5f7-1b6f34afac11" containerID="a4960334836f7e198fd176ea85315ee9f98088e691848d494e35542d764e7349" exitCode=0 Dec 04 13:11:03 crc kubenswrapper[4979]: I1204 13:11:03.546509 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-958bg" event={"ID":"5fa6e5e6-5309-489a-b5f7-1b6f34afac11","Type":"ContainerDied","Data":"a4960334836f7e198fd176ea85315ee9f98088e691848d494e35542d764e7349"} Dec 04 13:11:03 crc kubenswrapper[4979]: I1204 13:11:03.552042 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"f41cc400d512fbc84293fd7f01e9e77498ba3dd4183c76c64ed639782564147d"} Dec 04 13:11:04 crc kubenswrapper[4979]: I1204 13:11:04.212328 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" path="/var/lib/kubelet/pods/e9d44c50-a0aa-4aed-a2d0-ec957330b617/volumes" Dec 04 13:11:04 crc kubenswrapper[4979]: I1204 13:11:04.904826 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-958bg" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.075073 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-fernet-keys\") pod \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.075130 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-combined-ca-bundle\") pod \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.075278 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-credential-keys\") pod \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.075353 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-config-data\") pod \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.076020 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zktrz\" (UniqueName: \"kubernetes.io/projected/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-kube-api-access-zktrz\") pod \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.076093 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-scripts\") pod \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\" (UID: \"5fa6e5e6-5309-489a-b5f7-1b6f34afac11\") " Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.080668 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "5fa6e5e6-5309-489a-b5f7-1b6f34afac11" (UID: "5fa6e5e6-5309-489a-b5f7-1b6f34afac11"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.081056 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-kube-api-access-zktrz" (OuterVolumeSpecName: "kube-api-access-zktrz") pod "5fa6e5e6-5309-489a-b5f7-1b6f34afac11" (UID: "5fa6e5e6-5309-489a-b5f7-1b6f34afac11"). InnerVolumeSpecName "kube-api-access-zktrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.081247 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "5fa6e5e6-5309-489a-b5f7-1b6f34afac11" (UID: "5fa6e5e6-5309-489a-b5f7-1b6f34afac11"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.081454 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-scripts" (OuterVolumeSpecName: "scripts") pod "5fa6e5e6-5309-489a-b5f7-1b6f34afac11" (UID: "5fa6e5e6-5309-489a-b5f7-1b6f34afac11"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.099412 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-config-data" (OuterVolumeSpecName: "config-data") pod "5fa6e5e6-5309-489a-b5f7-1b6f34afac11" (UID: "5fa6e5e6-5309-489a-b5f7-1b6f34afac11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.102093 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5fa6e5e6-5309-489a-b5f7-1b6f34afac11" (UID: "5fa6e5e6-5309-489a-b5f7-1b6f34afac11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.177602 4979 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.177846 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.177936 4979 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.178063 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.178133 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zktrz\" (UniqueName: \"kubernetes.io/projected/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-kube-api-access-zktrz\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.178198 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5fa6e5e6-5309-489a-b5f7-1b6f34afac11-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.571802 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-958bg" event={"ID":"5fa6e5e6-5309-489a-b5f7-1b6f34afac11","Type":"ContainerDied","Data":"d568dd45ae2211eb0c50f0ab694d676dc60f83fad2b125ab477d40a2a526c448"} Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.572035 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d568dd45ae2211eb0c50f0ab694d676dc60f83fad2b125ab477d40a2a526c448" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.571893 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-958bg" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.674427 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-55d998f8fd-5tnfq"] Dec 04 13:11:05 crc kubenswrapper[4979]: E1204 13:11:05.675474 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerName="dnsmasq-dns" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.675597 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerName="dnsmasq-dns" Dec 04 13:11:05 crc kubenswrapper[4979]: E1204 13:11:05.675758 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa6e5e6-5309-489a-b5f7-1b6f34afac11" containerName="keystone-bootstrap" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.675851 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa6e5e6-5309-489a-b5f7-1b6f34afac11" containerName="keystone-bootstrap" Dec 04 13:11:05 crc kubenswrapper[4979]: E1204 13:11:05.675932 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerName="init" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.676014 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerName="init" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.676281 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa6e5e6-5309-489a-b5f7-1b6f34afac11" containerName="keystone-bootstrap" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.676404 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9d44c50-a0aa-4aed-a2d0-ec957330b617" containerName="dnsmasq-dns" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.677204 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.682066 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-4lfhw" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.682633 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.683024 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.683122 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.711137 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-55d998f8fd-5tnfq"] Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.789709 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-scripts\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.790102 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-combined-ca-bundle\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.790282 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-fernet-keys\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.790432 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-credential-keys\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.790583 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bvbp\" (UniqueName: \"kubernetes.io/projected/783f6a9e-4f95-4180-948e-122ec32a0deb-kube-api-access-6bvbp\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.790838 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-config-data\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.891871 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-fernet-keys\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.891951 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-credential-keys\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.891998 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bvbp\" (UniqueName: \"kubernetes.io/projected/783f6a9e-4f95-4180-948e-122ec32a0deb-kube-api-access-6bvbp\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.892077 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-config-data\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.892141 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-scripts\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.892174 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-combined-ca-bundle\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.898657 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-scripts\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.899519 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-config-data\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.900966 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-combined-ca-bundle\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.901648 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-fernet-keys\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.901717 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/783f6a9e-4f95-4180-948e-122ec32a0deb-credential-keys\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:05 crc kubenswrapper[4979]: I1204 13:11:05.914865 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bvbp\" (UniqueName: \"kubernetes.io/projected/783f6a9e-4f95-4180-948e-122ec32a0deb-kube-api-access-6bvbp\") pod \"keystone-55d998f8fd-5tnfq\" (UID: \"783f6a9e-4f95-4180-948e-122ec32a0deb\") " pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:05.999982 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.290242 4979 scope.go:117] "RemoveContainer" containerID="f6f5067562fb1e83b5940822e0b5dfed2b15298f6121aa32466c36cc1591520c" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.317878 4979 scope.go:117] "RemoveContainer" containerID="f0beb506bad144e6cc3bac4ccbcae612173f3b965f8d3f345a419c4bac2779fe" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.361882 4979 scope.go:117] "RemoveContainer" containerID="d3671dfc11e3788e515d961c21fbb488d7416ca55d5634f6eee5e24055d5ba67" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.381680 4979 scope.go:117] "RemoveContainer" containerID="1af760238713951c6db4cbd83f0171809978ab7502833f77d0b6f4836b22a786" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.402902 4979 scope.go:117] "RemoveContainer" containerID="8dedfa414fac3e373b31ea9d5fd3e57c8404a857c25566864a85c31dd85d7e91" Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.471075 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-55d998f8fd-5tnfq"] Dec 04 13:11:06 crc kubenswrapper[4979]: I1204 13:11:06.582047 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55d998f8fd-5tnfq" event={"ID":"783f6a9e-4f95-4180-948e-122ec32a0deb","Type":"ContainerStarted","Data":"f4915626097964c95b57d5ac1c53a1c3d6cbf63b9e3267ee10f6a74efeb23c8d"} Dec 04 13:11:07 crc kubenswrapper[4979]: I1204 13:11:07.593117 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-55d998f8fd-5tnfq" event={"ID":"783f6a9e-4f95-4180-948e-122ec32a0deb","Type":"ContainerStarted","Data":"d71a6ab27f6d063d15bc992ac5e8410b50d09bf45803c8a61330def121bfa63a"} Dec 04 13:11:08 crc kubenswrapper[4979]: I1204 13:11:08.601331 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:08 crc kubenswrapper[4979]: I1204 13:11:08.624060 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-55d998f8fd-5tnfq" podStartSLOduration=3.624036568 podStartE2EDuration="3.624036568s" podCreationTimestamp="2025-12-04 13:11:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:11:08.617770532 +0000 UTC m=+5292.892066336" watchObservedRunningTime="2025-12-04 13:11:08.624036568 +0000 UTC m=+5292.898332372" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.265861 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hlt69"] Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.268375 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.287822 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlt69"] Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.428425 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-catalog-content\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.428492 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9clrq\" (UniqueName: \"kubernetes.io/projected/98278c8f-c658-4656-a416-b932044eb824-kube-api-access-9clrq\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.428535 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-utilities\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.530620 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-catalog-content\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.530717 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9clrq\" (UniqueName: \"kubernetes.io/projected/98278c8f-c658-4656-a416-b932044eb824-kube-api-access-9clrq\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.530826 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-utilities\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.531398 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-catalog-content\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.531452 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-utilities\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.551388 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9clrq\" (UniqueName: \"kubernetes.io/projected/98278c8f-c658-4656-a416-b932044eb824-kube-api-access-9clrq\") pod \"redhat-marketplace-hlt69\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:32 crc kubenswrapper[4979]: I1204 13:11:32.608283 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:33 crc kubenswrapper[4979]: I1204 13:11:33.035489 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlt69"] Dec 04 13:11:33 crc kubenswrapper[4979]: I1204 13:11:33.808287 4979 generic.go:334] "Generic (PLEG): container finished" podID="98278c8f-c658-4656-a416-b932044eb824" containerID="16336e3ca17163a3fbf881a16754ee4d8d14ba0b765570d5535d9114d04bf29c" exitCode=0 Dec 04 13:11:33 crc kubenswrapper[4979]: I1204 13:11:33.808618 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlt69" event={"ID":"98278c8f-c658-4656-a416-b932044eb824","Type":"ContainerDied","Data":"16336e3ca17163a3fbf881a16754ee4d8d14ba0b765570d5535d9114d04bf29c"} Dec 04 13:11:33 crc kubenswrapper[4979]: I1204 13:11:33.808650 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlt69" event={"ID":"98278c8f-c658-4656-a416-b932044eb824","Type":"ContainerStarted","Data":"bc88e7305e47bbda22d79462e056cd9eb408d314892bcb8b8a136a6bc78c544d"} Dec 04 13:11:33 crc kubenswrapper[4979]: I1204 13:11:33.817181 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:11:35 crc kubenswrapper[4979]: I1204 13:11:35.825208 4979 generic.go:334] "Generic (PLEG): container finished" podID="98278c8f-c658-4656-a416-b932044eb824" containerID="9943beb63391072d403fb1455999aae801d667c2db3f8a86bcb2eee8b6fc7c5c" exitCode=0 Dec 04 13:11:35 crc kubenswrapper[4979]: I1204 13:11:35.825322 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlt69" event={"ID":"98278c8f-c658-4656-a416-b932044eb824","Type":"ContainerDied","Data":"9943beb63391072d403fb1455999aae801d667c2db3f8a86bcb2eee8b6fc7c5c"} Dec 04 13:11:36 crc kubenswrapper[4979]: I1204 13:11:36.836055 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlt69" event={"ID":"98278c8f-c658-4656-a416-b932044eb824","Type":"ContainerStarted","Data":"e6f41bf497bb9079ea2655a2287afa3a916e69e0b992f0c1df5e13e4bf13a3f4"} Dec 04 13:11:36 crc kubenswrapper[4979]: I1204 13:11:36.863652 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hlt69" podStartSLOduration=2.448055676 podStartE2EDuration="4.863623811s" podCreationTimestamp="2025-12-04 13:11:32 +0000 UTC" firstStartedPulling="2025-12-04 13:11:33.816892768 +0000 UTC m=+5318.091188572" lastFinishedPulling="2025-12-04 13:11:36.232460903 +0000 UTC m=+5320.506756707" observedRunningTime="2025-12-04 13:11:36.858085944 +0000 UTC m=+5321.132381748" watchObservedRunningTime="2025-12-04 13:11:36.863623811 +0000 UTC m=+5321.137919615" Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.555436 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-55d998f8fd-5tnfq" Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.914661 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.915767 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.920376 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-562vz" Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.920619 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.922176 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 04 13:11:37 crc kubenswrapper[4979]: I1204 13:11:37.933820 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.009556 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:38 crc kubenswrapper[4979]: E1204 13:11:38.010292 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[kube-api-access-hl6ks openstack-config openstack-config-secret], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/openstackclient" podUID="e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.019329 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl6ks\" (UniqueName: \"kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.019448 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config-secret\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.019521 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.042451 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.056382 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.057830 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.070614 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.121657 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config-secret\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.121765 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.121838 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.121885 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config-secret\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.121966 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpws4\" (UniqueName: \"kubernetes.io/projected/380f708e-bc42-431f-b675-af01d9c11ae8-kube-api-access-hpws4\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.121998 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl6ks\" (UniqueName: \"kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.123769 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: E1204 13:11:38.124685 4979 projected.go:194] Error preparing data for projected volume kube-api-access-hl6ks for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b) does not match the UID in record. The object might have been deleted and then recreated Dec 04 13:11:38 crc kubenswrapper[4979]: E1204 13:11:38.124766 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks podName:e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b nodeName:}" failed. No retries permitted until 2025-12-04 13:11:38.624742756 +0000 UTC m=+5322.899038560 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-hl6ks" (UniqueName: "kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks") pod "openstackclient" (UID: "e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b) does not match the UID in record. The object might have been deleted and then recreated Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.128558 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config-secret\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.223324 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.223397 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config-secret\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.223457 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpws4\" (UniqueName: \"kubernetes.io/projected/380f708e-bc42-431f-b675-af01d9c11ae8-kube-api-access-hpws4\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.224206 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.226696 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config-secret\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.241211 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpws4\" (UniqueName: \"kubernetes.io/projected/380f708e-bc42-431f-b675-af01d9c11ae8-kube-api-access-hpws4\") pod \"openstackclient\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.400840 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.629470 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl6ks\" (UniqueName: \"kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks\") pod \"openstackclient\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: E1204 13:11:38.631361 4979 projected.go:194] Error preparing data for projected volume kube-api-access-hl6ks for pod openstack/openstackclient: failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b) does not match the UID in record. The object might have been deleted and then recreated Dec 04 13:11:38 crc kubenswrapper[4979]: E1204 13:11:38.631420 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks podName:e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b nodeName:}" failed. No retries permitted until 2025-12-04 13:11:39.631402035 +0000 UTC m=+5323.905697839 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-hl6ks" (UniqueName: "kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks") pod "openstackclient" (UID: "e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b") : failed to fetch token: serviceaccounts "openstackclient-openstackclient" is forbidden: the UID in the bound object reference (e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b) does not match the UID in record. The object might have been deleted and then recreated Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.833852 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 13:11:38 crc kubenswrapper[4979]: W1204 13:11:38.835781 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod380f708e_bc42_431f_b675_af01d9c11ae8.slice/crio-be26e6d4ac808fd2a6bab250855983802dda0e2e463858c4e02b6909c10504e7 WatchSource:0}: Error finding container be26e6d4ac808fd2a6bab250855983802dda0e2e463858c4e02b6909c10504e7: Status 404 returned error can't find the container with id be26e6d4ac808fd2a6bab250855983802dda0e2e463858c4e02b6909c10504e7 Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.854241 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.854272 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"380f708e-bc42-431f-b675-af01d9c11ae8","Type":"ContainerStarted","Data":"be26e6d4ac808fd2a6bab250855983802dda0e2e463858c4e02b6909c10504e7"} Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.865265 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.869210 4979 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.934925 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config\") pod \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.935011 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config-secret\") pod \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\" (UID: \"e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b\") " Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.935342 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hl6ks\" (UniqueName: \"kubernetes.io/projected/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-kube-api-access-hl6ks\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.935478 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b" (UID: "e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:11:38 crc kubenswrapper[4979]: I1204 13:11:38.938510 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b" (UID: "e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:11:39 crc kubenswrapper[4979]: I1204 13:11:39.037161 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:39 crc kubenswrapper[4979]: I1204 13:11:39.037194 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:39 crc kubenswrapper[4979]: I1204 13:11:39.866811 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:11:39 crc kubenswrapper[4979]: I1204 13:11:39.866820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"380f708e-bc42-431f-b675-af01d9c11ae8","Type":"ContainerStarted","Data":"1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac"} Dec 04 13:11:39 crc kubenswrapper[4979]: I1204 13:11:39.886510 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=1.8864929799999999 podStartE2EDuration="1.88649298s" podCreationTimestamp="2025-12-04 13:11:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:11:39.883946142 +0000 UTC m=+5324.158241946" watchObservedRunningTime="2025-12-04 13:11:39.88649298 +0000 UTC m=+5324.160788784" Dec 04 13:11:39 crc kubenswrapper[4979]: I1204 13:11:39.890380 4979 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" Dec 04 13:11:40 crc kubenswrapper[4979]: I1204 13:11:40.210841 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b" path="/var/lib/kubelet/pods/e9c9b33a-ef56-49c4-b3fd-5d84c9853f5b/volumes" Dec 04 13:11:42 crc kubenswrapper[4979]: I1204 13:11:42.608534 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:42 crc kubenswrapper[4979]: I1204 13:11:42.609162 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:42 crc kubenswrapper[4979]: I1204 13:11:42.661156 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:42 crc kubenswrapper[4979]: I1204 13:11:42.941781 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:43 crc kubenswrapper[4979]: I1204 13:11:43.000273 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlt69"] Dec 04 13:11:44 crc kubenswrapper[4979]: I1204 13:11:44.909337 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hlt69" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="registry-server" containerID="cri-o://e6f41bf497bb9079ea2655a2287afa3a916e69e0b992f0c1df5e13e4bf13a3f4" gracePeriod=2 Dec 04 13:11:45 crc kubenswrapper[4979]: I1204 13:11:45.919083 4979 generic.go:334] "Generic (PLEG): container finished" podID="98278c8f-c658-4656-a416-b932044eb824" containerID="e6f41bf497bb9079ea2655a2287afa3a916e69e0b992f0c1df5e13e4bf13a3f4" exitCode=0 Dec 04 13:11:45 crc kubenswrapper[4979]: I1204 13:11:45.919177 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlt69" event={"ID":"98278c8f-c658-4656-a416-b932044eb824","Type":"ContainerDied","Data":"e6f41bf497bb9079ea2655a2287afa3a916e69e0b992f0c1df5e13e4bf13a3f4"} Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.433039 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.549629 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-utilities\") pod \"98278c8f-c658-4656-a416-b932044eb824\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.549703 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9clrq\" (UniqueName: \"kubernetes.io/projected/98278c8f-c658-4656-a416-b932044eb824-kube-api-access-9clrq\") pod \"98278c8f-c658-4656-a416-b932044eb824\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.549791 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-catalog-content\") pod \"98278c8f-c658-4656-a416-b932044eb824\" (UID: \"98278c8f-c658-4656-a416-b932044eb824\") " Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.550623 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-utilities" (OuterVolumeSpecName: "utilities") pod "98278c8f-c658-4656-a416-b932044eb824" (UID: "98278c8f-c658-4656-a416-b932044eb824"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.557760 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98278c8f-c658-4656-a416-b932044eb824-kube-api-access-9clrq" (OuterVolumeSpecName: "kube-api-access-9clrq") pod "98278c8f-c658-4656-a416-b932044eb824" (UID: "98278c8f-c658-4656-a416-b932044eb824"). InnerVolumeSpecName "kube-api-access-9clrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.568051 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "98278c8f-c658-4656-a416-b932044eb824" (UID: "98278c8f-c658-4656-a416-b932044eb824"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.651813 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.651847 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9clrq\" (UniqueName: \"kubernetes.io/projected/98278c8f-c658-4656-a416-b932044eb824-kube-api-access-9clrq\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.651860 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/98278c8f-c658-4656-a416-b932044eb824-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.928226 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hlt69" event={"ID":"98278c8f-c658-4656-a416-b932044eb824","Type":"ContainerDied","Data":"bc88e7305e47bbda22d79462e056cd9eb408d314892bcb8b8a136a6bc78c544d"} Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.928277 4979 scope.go:117] "RemoveContainer" containerID="e6f41bf497bb9079ea2655a2287afa3a916e69e0b992f0c1df5e13e4bf13a3f4" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.928843 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hlt69" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.957707 4979 scope.go:117] "RemoveContainer" containerID="9943beb63391072d403fb1455999aae801d667c2db3f8a86bcb2eee8b6fc7c5c" Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.962189 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlt69"] Dec 04 13:11:46 crc kubenswrapper[4979]: I1204 13:11:46.968912 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hlt69"] Dec 04 13:11:47 crc kubenswrapper[4979]: I1204 13:11:47.001140 4979 scope.go:117] "RemoveContainer" containerID="16336e3ca17163a3fbf881a16754ee4d8d14ba0b765570d5535d9114d04bf29c" Dec 04 13:11:48 crc kubenswrapper[4979]: I1204 13:11:48.209820 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98278c8f-c658-4656-a416-b932044eb824" path="/var/lib/kubelet/pods/98278c8f-c658-4656-a416-b932044eb824/volumes" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.622031 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-8snsg"] Dec 04 13:12:56 crc kubenswrapper[4979]: E1204 13:12:56.623026 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="extract-utilities" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.623042 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="extract-utilities" Dec 04 13:12:56 crc kubenswrapper[4979]: E1204 13:12:56.623066 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="registry-server" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.623074 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="registry-server" Dec 04 13:12:56 crc kubenswrapper[4979]: E1204 13:12:56.623108 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="extract-content" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.623119 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="extract-content" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.623353 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="98278c8f-c658-4656-a416-b932044eb824" containerName="registry-server" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.623999 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.634324 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-bf3c-account-create-update-k78d4"] Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.635973 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.637790 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.651407 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bf3c-account-create-update-k78d4"] Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.662593 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8snsg"] Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.670745 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2874\" (UniqueName: \"kubernetes.io/projected/a4bd880a-4a20-4038-a98a-7fc2c170c739-kube-api-access-t2874\") pod \"barbican-db-create-8snsg\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.670797 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d27ea1-b55f-4072-af5d-7dfa136317a2-operator-scripts\") pod \"barbican-bf3c-account-create-update-k78d4\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.670827 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4bd880a-4a20-4038-a98a-7fc2c170c739-operator-scripts\") pod \"barbican-db-create-8snsg\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.670915 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdmzg\" (UniqueName: \"kubernetes.io/projected/35d27ea1-b55f-4072-af5d-7dfa136317a2-kube-api-access-vdmzg\") pod \"barbican-bf3c-account-create-update-k78d4\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.773047 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2874\" (UniqueName: \"kubernetes.io/projected/a4bd880a-4a20-4038-a98a-7fc2c170c739-kube-api-access-t2874\") pod \"barbican-db-create-8snsg\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.773145 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d27ea1-b55f-4072-af5d-7dfa136317a2-operator-scripts\") pod \"barbican-bf3c-account-create-update-k78d4\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.773190 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4bd880a-4a20-4038-a98a-7fc2c170c739-operator-scripts\") pod \"barbican-db-create-8snsg\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.773290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdmzg\" (UniqueName: \"kubernetes.io/projected/35d27ea1-b55f-4072-af5d-7dfa136317a2-kube-api-access-vdmzg\") pod \"barbican-bf3c-account-create-update-k78d4\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.774213 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4bd880a-4a20-4038-a98a-7fc2c170c739-operator-scripts\") pod \"barbican-db-create-8snsg\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.774242 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d27ea1-b55f-4072-af5d-7dfa136317a2-operator-scripts\") pod \"barbican-bf3c-account-create-update-k78d4\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.796472 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdmzg\" (UniqueName: \"kubernetes.io/projected/35d27ea1-b55f-4072-af5d-7dfa136317a2-kube-api-access-vdmzg\") pod \"barbican-bf3c-account-create-update-k78d4\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.796758 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2874\" (UniqueName: \"kubernetes.io/projected/a4bd880a-4a20-4038-a98a-7fc2c170c739-kube-api-access-t2874\") pod \"barbican-db-create-8snsg\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.963009 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8snsg" Dec 04 13:12:56 crc kubenswrapper[4979]: I1204 13:12:56.974545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:57 crc kubenswrapper[4979]: I1204 13:12:57.230015 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-8snsg"] Dec 04 13:12:57 crc kubenswrapper[4979]: I1204 13:12:57.512400 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bf3c-account-create-update-k78d4"] Dec 04 13:12:57 crc kubenswrapper[4979]: W1204 13:12:57.513770 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35d27ea1_b55f_4072_af5d_7dfa136317a2.slice/crio-33ab4390059fe02c33b628c8ef5c0d895fdacaf64fc2fb2f9c2e4e4c65861134 WatchSource:0}: Error finding container 33ab4390059fe02c33b628c8ef5c0d895fdacaf64fc2fb2f9c2e4e4c65861134: Status 404 returned error can't find the container with id 33ab4390059fe02c33b628c8ef5c0d895fdacaf64fc2fb2f9c2e4e4c65861134 Dec 04 13:12:57 crc kubenswrapper[4979]: I1204 13:12:57.514812 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8snsg" event={"ID":"a4bd880a-4a20-4038-a98a-7fc2c170c739","Type":"ContainerStarted","Data":"d6e01d06690e670f9a5f9acfd54c418b55bf49e11f41246df5d8a0706f012ccf"} Dec 04 13:12:57 crc kubenswrapper[4979]: I1204 13:12:57.514863 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8snsg" event={"ID":"a4bd880a-4a20-4038-a98a-7fc2c170c739","Type":"ContainerStarted","Data":"757abf15042bf18a8f49676c741e411a416d02aac07c541430b8bc15fd752244"} Dec 04 13:12:57 crc kubenswrapper[4979]: I1204 13:12:57.539835 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-8snsg" podStartSLOduration=1.539813071 podStartE2EDuration="1.539813071s" podCreationTimestamp="2025-12-04 13:12:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:12:57.53449195 +0000 UTC m=+5401.808787754" watchObservedRunningTime="2025-12-04 13:12:57.539813071 +0000 UTC m=+5401.814108875" Dec 04 13:12:58 crc kubenswrapper[4979]: I1204 13:12:58.525127 4979 generic.go:334] "Generic (PLEG): container finished" podID="35d27ea1-b55f-4072-af5d-7dfa136317a2" containerID="802a051e7514b076257335eed5d687cdda129a9f5c03f81043742d58e215b285" exitCode=0 Dec 04 13:12:58 crc kubenswrapper[4979]: I1204 13:12:58.525266 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf3c-account-create-update-k78d4" event={"ID":"35d27ea1-b55f-4072-af5d-7dfa136317a2","Type":"ContainerDied","Data":"802a051e7514b076257335eed5d687cdda129a9f5c03f81043742d58e215b285"} Dec 04 13:12:58 crc kubenswrapper[4979]: I1204 13:12:58.526552 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf3c-account-create-update-k78d4" event={"ID":"35d27ea1-b55f-4072-af5d-7dfa136317a2","Type":"ContainerStarted","Data":"33ab4390059fe02c33b628c8ef5c0d895fdacaf64fc2fb2f9c2e4e4c65861134"} Dec 04 13:12:58 crc kubenswrapper[4979]: I1204 13:12:58.528555 4979 generic.go:334] "Generic (PLEG): container finished" podID="a4bd880a-4a20-4038-a98a-7fc2c170c739" containerID="d6e01d06690e670f9a5f9acfd54c418b55bf49e11f41246df5d8a0706f012ccf" exitCode=0 Dec 04 13:12:58 crc kubenswrapper[4979]: I1204 13:12:58.528604 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8snsg" event={"ID":"a4bd880a-4a20-4038-a98a-7fc2c170c739","Type":"ContainerDied","Data":"d6e01d06690e670f9a5f9acfd54c418b55bf49e11f41246df5d8a0706f012ccf"} Dec 04 13:12:59 crc kubenswrapper[4979]: I1204 13:12:59.940670 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:12:59 crc kubenswrapper[4979]: I1204 13:12:59.949827 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8snsg" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.033055 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2874\" (UniqueName: \"kubernetes.io/projected/a4bd880a-4a20-4038-a98a-7fc2c170c739-kube-api-access-t2874\") pod \"a4bd880a-4a20-4038-a98a-7fc2c170c739\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.033164 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4bd880a-4a20-4038-a98a-7fc2c170c739-operator-scripts\") pod \"a4bd880a-4a20-4038-a98a-7fc2c170c739\" (UID: \"a4bd880a-4a20-4038-a98a-7fc2c170c739\") " Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.033227 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d27ea1-b55f-4072-af5d-7dfa136317a2-operator-scripts\") pod \"35d27ea1-b55f-4072-af5d-7dfa136317a2\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.033277 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vdmzg\" (UniqueName: \"kubernetes.io/projected/35d27ea1-b55f-4072-af5d-7dfa136317a2-kube-api-access-vdmzg\") pod \"35d27ea1-b55f-4072-af5d-7dfa136317a2\" (UID: \"35d27ea1-b55f-4072-af5d-7dfa136317a2\") " Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.034096 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35d27ea1-b55f-4072-af5d-7dfa136317a2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35d27ea1-b55f-4072-af5d-7dfa136317a2" (UID: "35d27ea1-b55f-4072-af5d-7dfa136317a2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.034875 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4bd880a-4a20-4038-a98a-7fc2c170c739-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a4bd880a-4a20-4038-a98a-7fc2c170c739" (UID: "a4bd880a-4a20-4038-a98a-7fc2c170c739"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.038859 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4bd880a-4a20-4038-a98a-7fc2c170c739-kube-api-access-t2874" (OuterVolumeSpecName: "kube-api-access-t2874") pod "a4bd880a-4a20-4038-a98a-7fc2c170c739" (UID: "a4bd880a-4a20-4038-a98a-7fc2c170c739"). InnerVolumeSpecName "kube-api-access-t2874". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.039196 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d27ea1-b55f-4072-af5d-7dfa136317a2-kube-api-access-vdmzg" (OuterVolumeSpecName: "kube-api-access-vdmzg") pod "35d27ea1-b55f-4072-af5d-7dfa136317a2" (UID: "35d27ea1-b55f-4072-af5d-7dfa136317a2"). InnerVolumeSpecName "kube-api-access-vdmzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.134869 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2874\" (UniqueName: \"kubernetes.io/projected/a4bd880a-4a20-4038-a98a-7fc2c170c739-kube-api-access-t2874\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.134920 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a4bd880a-4a20-4038-a98a-7fc2c170c739-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.134935 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35d27ea1-b55f-4072-af5d-7dfa136317a2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.134948 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vdmzg\" (UniqueName: \"kubernetes.io/projected/35d27ea1-b55f-4072-af5d-7dfa136317a2-kube-api-access-vdmzg\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.547932 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-8snsg" event={"ID":"a4bd880a-4a20-4038-a98a-7fc2c170c739","Type":"ContainerDied","Data":"757abf15042bf18a8f49676c741e411a416d02aac07c541430b8bc15fd752244"} Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.548221 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="757abf15042bf18a8f49676c741e411a416d02aac07c541430b8bc15fd752244" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.547944 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-8snsg" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.549672 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bf3c-account-create-update-k78d4" event={"ID":"35d27ea1-b55f-4072-af5d-7dfa136317a2","Type":"ContainerDied","Data":"33ab4390059fe02c33b628c8ef5c0d895fdacaf64fc2fb2f9c2e4e4c65861134"} Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.549707 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33ab4390059fe02c33b628c8ef5c0d895fdacaf64fc2fb2f9c2e4e4c65861134" Dec 04 13:13:00 crc kubenswrapper[4979]: I1204 13:13:00.549711 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bf3c-account-create-update-k78d4" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.814226 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-srlgj"] Dec 04 13:13:01 crc kubenswrapper[4979]: E1204 13:13:01.814629 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d27ea1-b55f-4072-af5d-7dfa136317a2" containerName="mariadb-account-create-update" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.814650 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d27ea1-b55f-4072-af5d-7dfa136317a2" containerName="mariadb-account-create-update" Dec 04 13:13:01 crc kubenswrapper[4979]: E1204 13:13:01.814680 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4bd880a-4a20-4038-a98a-7fc2c170c739" containerName="mariadb-database-create" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.814689 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4bd880a-4a20-4038-a98a-7fc2c170c739" containerName="mariadb-database-create" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.814888 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4bd880a-4a20-4038-a98a-7fc2c170c739" containerName="mariadb-database-create" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.814919 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d27ea1-b55f-4072-af5d-7dfa136317a2" containerName="mariadb-account-create-update" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.815692 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.818247 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.820574 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ml67d" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.832715 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-srlgj"] Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.864088 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-combined-ca-bundle\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.864160 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwjmn\" (UniqueName: \"kubernetes.io/projected/39e0503d-f5da-4fe8-8488-13e9e2834b8f-kube-api-access-dwjmn\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.864563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-db-sync-config-data\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.966361 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-combined-ca-bundle\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.966620 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwjmn\" (UniqueName: \"kubernetes.io/projected/39e0503d-f5da-4fe8-8488-13e9e2834b8f-kube-api-access-dwjmn\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.966908 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-db-sync-config-data\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.982989 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-db-sync-config-data\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.985627 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-combined-ca-bundle\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:01 crc kubenswrapper[4979]: I1204 13:13:01.986411 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwjmn\" (UniqueName: \"kubernetes.io/projected/39e0503d-f5da-4fe8-8488-13e9e2834b8f-kube-api-access-dwjmn\") pod \"barbican-db-sync-srlgj\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:02 crc kubenswrapper[4979]: I1204 13:13:02.133642 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:02 crc kubenswrapper[4979]: I1204 13:13:02.582196 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-srlgj"] Dec 04 13:13:02 crc kubenswrapper[4979]: W1204 13:13:02.587524 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39e0503d_f5da_4fe8_8488_13e9e2834b8f.slice/crio-e578a47f27ece0de188ab600e2247f1793d61e18cdce7b3780d4e7b02457486c WatchSource:0}: Error finding container e578a47f27ece0de188ab600e2247f1793d61e18cdce7b3780d4e7b02457486c: Status 404 returned error can't find the container with id e578a47f27ece0de188ab600e2247f1793d61e18cdce7b3780d4e7b02457486c Dec 04 13:13:03 crc kubenswrapper[4979]: I1204 13:13:03.585688 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srlgj" event={"ID":"39e0503d-f5da-4fe8-8488-13e9e2834b8f","Type":"ContainerStarted","Data":"dc559ab693d8c1733e9c4b638a86fe0dea3a042d875ae00c7203ba142adec4c0"} Dec 04 13:13:03 crc kubenswrapper[4979]: I1204 13:13:03.588572 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srlgj" event={"ID":"39e0503d-f5da-4fe8-8488-13e9e2834b8f","Type":"ContainerStarted","Data":"e578a47f27ece0de188ab600e2247f1793d61e18cdce7b3780d4e7b02457486c"} Dec 04 13:13:03 crc kubenswrapper[4979]: I1204 13:13:03.640474 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-srlgj" podStartSLOduration=2.640453899 podStartE2EDuration="2.640453899s" podCreationTimestamp="2025-12-04 13:13:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:03.61156443 +0000 UTC m=+5407.885860254" watchObservedRunningTime="2025-12-04 13:13:03.640453899 +0000 UTC m=+5407.914749703" Dec 04 13:13:04 crc kubenswrapper[4979]: I1204 13:13:04.597148 4979 generic.go:334] "Generic (PLEG): container finished" podID="39e0503d-f5da-4fe8-8488-13e9e2834b8f" containerID="dc559ab693d8c1733e9c4b638a86fe0dea3a042d875ae00c7203ba142adec4c0" exitCode=0 Dec 04 13:13:04 crc kubenswrapper[4979]: I1204 13:13:04.597233 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srlgj" event={"ID":"39e0503d-f5da-4fe8-8488-13e9e2834b8f","Type":"ContainerDied","Data":"dc559ab693d8c1733e9c4b638a86fe0dea3a042d875ae00c7203ba142adec4c0"} Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.902525 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.931052 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-combined-ca-bundle\") pod \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.931152 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwjmn\" (UniqueName: \"kubernetes.io/projected/39e0503d-f5da-4fe8-8488-13e9e2834b8f-kube-api-access-dwjmn\") pod \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.931343 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-db-sync-config-data\") pod \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\" (UID: \"39e0503d-f5da-4fe8-8488-13e9e2834b8f\") " Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.936265 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "39e0503d-f5da-4fe8-8488-13e9e2834b8f" (UID: "39e0503d-f5da-4fe8-8488-13e9e2834b8f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.936941 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39e0503d-f5da-4fe8-8488-13e9e2834b8f-kube-api-access-dwjmn" (OuterVolumeSpecName: "kube-api-access-dwjmn") pod "39e0503d-f5da-4fe8-8488-13e9e2834b8f" (UID: "39e0503d-f5da-4fe8-8488-13e9e2834b8f"). InnerVolumeSpecName "kube-api-access-dwjmn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:05 crc kubenswrapper[4979]: I1204 13:13:05.957141 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39e0503d-f5da-4fe8-8488-13e9e2834b8f" (UID: "39e0503d-f5da-4fe8-8488-13e9e2834b8f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:13:06 crc kubenswrapper[4979]: I1204 13:13:06.033042 4979 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:06 crc kubenswrapper[4979]: I1204 13:13:06.033092 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39e0503d-f5da-4fe8-8488-13e9e2834b8f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:06 crc kubenswrapper[4979]: I1204 13:13:06.033105 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwjmn\" (UniqueName: \"kubernetes.io/projected/39e0503d-f5da-4fe8-8488-13e9e2834b8f-kube-api-access-dwjmn\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:06 crc kubenswrapper[4979]: I1204 13:13:06.615673 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-srlgj" event={"ID":"39e0503d-f5da-4fe8-8488-13e9e2834b8f","Type":"ContainerDied","Data":"e578a47f27ece0de188ab600e2247f1793d61e18cdce7b3780d4e7b02457486c"} Dec 04 13:13:06 crc kubenswrapper[4979]: I1204 13:13:06.615717 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-srlgj" Dec 04 13:13:06 crc kubenswrapper[4979]: I1204 13:13:06.615725 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e578a47f27ece0de188ab600e2247f1793d61e18cdce7b3780d4e7b02457486c" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.156286 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6bb568464f-gwp45"] Dec 04 13:13:07 crc kubenswrapper[4979]: E1204 13:13:07.157095 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39e0503d-f5da-4fe8-8488-13e9e2834b8f" containerName="barbican-db-sync" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.157114 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="39e0503d-f5da-4fe8-8488-13e9e2834b8f" containerName="barbican-db-sync" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.157348 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="39e0503d-f5da-4fe8-8488-13e9e2834b8f" containerName="barbican-db-sync" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.158481 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.169433 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.169710 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-ml67d" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.169995 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.174093 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-68445fcf8b-jnsbd"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.175842 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.179913 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.198385 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68445fcf8b-jnsbd"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.210742 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6bb568464f-gwp45"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250708 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-config-data-custom\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250805 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-config-data\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250829 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4dxz\" (UniqueName: \"kubernetes.io/projected/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-kube-api-access-s4dxz\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250854 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-config-data-custom\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250886 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-logs\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250902 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-config-data\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250949 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-combined-ca-bundle\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.250973 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-logs\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.251003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbpnw\" (UniqueName: \"kubernetes.io/projected/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-kube-api-access-zbpnw\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.251036 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-combined-ca-bundle\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.318604 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6d95f7-bxwbv"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.327741 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.329581 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6d95f7-bxwbv"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352425 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-config-data-custom\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352521 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-dns-svc\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352566 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-logs\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352585 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-config-data\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352613 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-config\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352656 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352687 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-combined-ca-bundle\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352714 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlvfd\" (UniqueName: \"kubernetes.io/projected/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-kube-api-access-tlvfd\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352737 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-logs\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352800 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbpnw\" (UniqueName: \"kubernetes.io/projected/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-kube-api-access-zbpnw\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352845 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-combined-ca-bundle\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352894 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-config-data-custom\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352950 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-config-data\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352970 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4dxz\" (UniqueName: \"kubernetes.io/projected/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-kube-api-access-s4dxz\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.352991 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.353953 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-logs\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.354286 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-logs\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.358494 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-config-data-custom\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.359870 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-config-data-custom\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.363926 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-config-data\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.364105 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-combined-ca-bundle\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.370962 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-config-data\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.379995 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbpnw\" (UniqueName: \"kubernetes.io/projected/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-kube-api-access-zbpnw\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.389608 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4dxz\" (UniqueName: \"kubernetes.io/projected/3d82b744-e656-40b6-bb9c-c7ffc8c007cd-kube-api-access-s4dxz\") pod \"barbican-keystone-listener-68445fcf8b-jnsbd\" (UID: \"3d82b744-e656-40b6-bb9c-c7ffc8c007cd\") " pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.402086 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3-combined-ca-bundle\") pod \"barbican-worker-6bb568464f-gwp45\" (UID: \"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3\") " pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.419892 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5c694d4454-2bzkv"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.421878 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.424950 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.430831 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c694d4454-2bzkv"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454242 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454326 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlvfd\" (UniqueName: \"kubernetes.io/projected/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-kube-api-access-tlvfd\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454546 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-config-data-custom\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454582 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-combined-ca-bundle\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454613 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-config-data\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454646 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvpl5\" (UniqueName: \"kubernetes.io/projected/a014ea0c-872a-4cd9-a428-596d863e2651-kube-api-access-zvpl5\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454697 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454724 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a014ea0c-872a-4cd9-a428-596d863e2651-logs\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454767 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-dns-svc\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.454798 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-config\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.455555 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.455685 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-config\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.456326 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.456871 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-dns-svc\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.476978 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlvfd\" (UniqueName: \"kubernetes.io/projected/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-kube-api-access-tlvfd\") pod \"dnsmasq-dns-8c6d95f7-bxwbv\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.492940 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6bb568464f-gwp45" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.521367 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.556226 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-config-data-custom\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.556279 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-combined-ca-bundle\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.556337 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-config-data\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.556376 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvpl5\" (UniqueName: \"kubernetes.io/projected/a014ea0c-872a-4cd9-a428-596d863e2651-kube-api-access-zvpl5\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.556438 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a014ea0c-872a-4cd9-a428-596d863e2651-logs\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.558471 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a014ea0c-872a-4cd9-a428-596d863e2651-logs\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.566562 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-config-data-custom\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.567801 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-config-data\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.568088 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a014ea0c-872a-4cd9-a428-596d863e2651-combined-ca-bundle\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.578198 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvpl5\" (UniqueName: \"kubernetes.io/projected/a014ea0c-872a-4cd9-a428-596d863e2651-kube-api-access-zvpl5\") pod \"barbican-api-5c694d4454-2bzkv\" (UID: \"a014ea0c-872a-4cd9-a428-596d863e2651\") " pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.626045 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.774444 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.959497 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5c694d4454-2bzkv"] Dec 04 13:13:07 crc kubenswrapper[4979]: I1204 13:13:07.999349 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6bb568464f-gwp45"] Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.052918 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-68445fcf8b-jnsbd"] Dec 04 13:13:08 crc kubenswrapper[4979]: W1204 13:13:08.069882 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d82b744_e656_40b6_bb9c_c7ffc8c007cd.slice/crio-5295a667fd26aa596f9f8bbbecd367337cc4ace81517bb567281beb3b2d18dbf WatchSource:0}: Error finding container 5295a667fd26aa596f9f8bbbecd367337cc4ace81517bb567281beb3b2d18dbf: Status 404 returned error can't find the container with id 5295a667fd26aa596f9f8bbbecd367337cc4ace81517bb567281beb3b2d18dbf Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.267551 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6d95f7-bxwbv"] Dec 04 13:13:08 crc kubenswrapper[4979]: W1204 13:13:08.285387 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7e6b5a2_0b10_449e_ae95_11dfebf8f956.slice/crio-91bdc0da5348a3b0b7366b876e7d76b651c0ac25a5adb1604b98ca3d2de92578 WatchSource:0}: Error finding container 91bdc0da5348a3b0b7366b876e7d76b651c0ac25a5adb1604b98ca3d2de92578: Status 404 returned error can't find the container with id 91bdc0da5348a3b0b7366b876e7d76b651c0ac25a5adb1604b98ca3d2de92578 Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.642936 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bb568464f-gwp45" event={"ID":"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3","Type":"ContainerStarted","Data":"aef27bdb6e7ba8c7306f2f108fe14e962ba8a1ae3cb3894f81e0028c5f370f05"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.643728 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bb568464f-gwp45" event={"ID":"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3","Type":"ContainerStarted","Data":"c45b868044af46be3bc97755abdaf335e6243f344e396e76c0e0ff93c0da361c"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.643751 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6bb568464f-gwp45" event={"ID":"9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3","Type":"ContainerStarted","Data":"ff3ce798d459fc1aa3c1f7805edca8d0a7041c9083f3dd34727e6ce6ec2dadf0"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.647818 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c694d4454-2bzkv" event={"ID":"a014ea0c-872a-4cd9-a428-596d863e2651","Type":"ContainerStarted","Data":"0990df789c392ac0285ca2a11dd10fe70f41bee3c4a50ddba009ce8152e91da7"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.647883 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c694d4454-2bzkv" event={"ID":"a014ea0c-872a-4cd9-a428-596d863e2651","Type":"ContainerStarted","Data":"0846e72901703c4e6ca5169c9bb88906b97e61959aa39ec5bd7af5c5eaaa0319"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.647904 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5c694d4454-2bzkv" event={"ID":"a014ea0c-872a-4cd9-a428-596d863e2651","Type":"ContainerStarted","Data":"527728489df772b1876dd4ba5c829d5304fc27e0cf642d209e214bbd2d0034ef"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.647961 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.647990 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.650589 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" event={"ID":"d7e6b5a2-0b10-449e-ae95-11dfebf8f956","Type":"ContainerStarted","Data":"149fcdd085c77a149377a9dfeb600153259be2f67df4f56c2c405b353a2f8c57"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.650628 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" event={"ID":"d7e6b5a2-0b10-449e-ae95-11dfebf8f956","Type":"ContainerStarted","Data":"91bdc0da5348a3b0b7366b876e7d76b651c0ac25a5adb1604b98ca3d2de92578"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.654104 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" event={"ID":"3d82b744-e656-40b6-bb9c-c7ffc8c007cd","Type":"ContainerStarted","Data":"5487d70c18ce3dd2019e0d9d264d379fa8705ca6ea8a2eb38b86fc8c22a81be7"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.654131 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" event={"ID":"3d82b744-e656-40b6-bb9c-c7ffc8c007cd","Type":"ContainerStarted","Data":"14559f2339a5a4d5209b3fc49b86a423c7a6960bbd4c8b2221c04d0c9f5b42d3"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.654142 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" event={"ID":"3d82b744-e656-40b6-bb9c-c7ffc8c007cd","Type":"ContainerStarted","Data":"5295a667fd26aa596f9f8bbbecd367337cc4ace81517bb567281beb3b2d18dbf"} Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.670769 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6bb568464f-gwp45" podStartSLOduration=1.670743942 podStartE2EDuration="1.670743942s" podCreationTimestamp="2025-12-04 13:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:08.661287511 +0000 UTC m=+5412.935583335" watchObservedRunningTime="2025-12-04 13:13:08.670743942 +0000 UTC m=+5412.945039756" Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.721103 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-68445fcf8b-jnsbd" podStartSLOduration=1.7210800499999999 podStartE2EDuration="1.72108005s" podCreationTimestamp="2025-12-04 13:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:08.717902786 +0000 UTC m=+5412.992198600" watchObservedRunningTime="2025-12-04 13:13:08.72108005 +0000 UTC m=+5412.995375854" Dec 04 13:13:08 crc kubenswrapper[4979]: I1204 13:13:08.753293 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5c694d4454-2bzkv" podStartSLOduration=1.7532754160000001 podStartE2EDuration="1.753275416s" podCreationTimestamp="2025-12-04 13:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:08.751487168 +0000 UTC m=+5413.025782972" watchObservedRunningTime="2025-12-04 13:13:08.753275416 +0000 UTC m=+5413.027571220" Dec 04 13:13:09 crc kubenswrapper[4979]: I1204 13:13:09.664433 4979 generic.go:334] "Generic (PLEG): container finished" podID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerID="149fcdd085c77a149377a9dfeb600153259be2f67df4f56c2c405b353a2f8c57" exitCode=0 Dec 04 13:13:09 crc kubenswrapper[4979]: I1204 13:13:09.664511 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" event={"ID":"d7e6b5a2-0b10-449e-ae95-11dfebf8f956","Type":"ContainerDied","Data":"149fcdd085c77a149377a9dfeb600153259be2f67df4f56c2c405b353a2f8c57"} Dec 04 13:13:09 crc kubenswrapper[4979]: I1204 13:13:09.664857 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" event={"ID":"d7e6b5a2-0b10-449e-ae95-11dfebf8f956","Type":"ContainerStarted","Data":"7064714334f098d81b367567cebbb47b023b8e2ff638bdaf9537aa2d9dddfaa3"} Dec 04 13:13:09 crc kubenswrapper[4979]: I1204 13:13:09.665253 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:09 crc kubenswrapper[4979]: I1204 13:13:09.683800 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" podStartSLOduration=2.683770102 podStartE2EDuration="2.683770102s" podCreationTimestamp="2025-12-04 13:13:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:09.682667743 +0000 UTC m=+5413.956963557" watchObservedRunningTime="2025-12-04 13:13:09.683770102 +0000 UTC m=+5413.958065916" Dec 04 13:13:17 crc kubenswrapper[4979]: I1204 13:13:17.777548 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:17 crc kubenswrapper[4979]: I1204 13:13:17.850944 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-96578f89-dn7t8"] Dec 04 13:13:17 crc kubenswrapper[4979]: I1204 13:13:17.851469 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-96578f89-dn7t8" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerName="dnsmasq-dns" containerID="cri-o://088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5" gracePeriod=10 Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.414138 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.550117 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-dns-svc\") pod \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.550173 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-nb\") pod \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.550220 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-sb\") pod \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.550344 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9gvt\" (UniqueName: \"kubernetes.io/projected/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-kube-api-access-b9gvt\") pod \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.550382 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-config\") pod \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\" (UID: \"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee\") " Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.564278 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-kube-api-access-b9gvt" (OuterVolumeSpecName: "kube-api-access-b9gvt") pod "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" (UID: "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee"). InnerVolumeSpecName "kube-api-access-b9gvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.598268 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" (UID: "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.601284 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-config" (OuterVolumeSpecName: "config") pod "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" (UID: "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.608787 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" (UID: "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.612719 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" (UID: "7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.652713 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9gvt\" (UniqueName: \"kubernetes.io/projected/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-kube-api-access-b9gvt\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.652746 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.652785 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.652800 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.652815 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.742969 4979 generic.go:334] "Generic (PLEG): container finished" podID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerID="088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5" exitCode=0 Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.743014 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-96578f89-dn7t8" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.743047 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96578f89-dn7t8" event={"ID":"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee","Type":"ContainerDied","Data":"088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5"} Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.743091 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-96578f89-dn7t8" event={"ID":"7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee","Type":"ContainerDied","Data":"3f01fc88f0e647edac4410bdba29d0c1e1c8573aecc958fa3554b59c77e81184"} Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.743139 4979 scope.go:117] "RemoveContainer" containerID="088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.767065 4979 scope.go:117] "RemoveContainer" containerID="45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.800646 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-96578f89-dn7t8"] Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.804901 4979 scope.go:117] "RemoveContainer" containerID="088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5" Dec 04 13:13:18 crc kubenswrapper[4979]: E1204 13:13:18.805648 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5\": container with ID starting with 088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5 not found: ID does not exist" containerID="088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.805698 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5"} err="failed to get container status \"088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5\": rpc error: code = NotFound desc = could not find container \"088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5\": container with ID starting with 088c09c4b7746c41db43de24f1fd32fb1b4f91ee9ad1acea53ac3de1ea3f71f5 not found: ID does not exist" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.805728 4979 scope.go:117] "RemoveContainer" containerID="45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580" Dec 04 13:13:18 crc kubenswrapper[4979]: E1204 13:13:18.806204 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580\": container with ID starting with 45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580 not found: ID does not exist" containerID="45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.806242 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580"} err="failed to get container status \"45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580\": rpc error: code = NotFound desc = could not find container \"45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580\": container with ID starting with 45febfc190245aadd43b833120f557fe82b28005b89683dc72da830750ca8580 not found: ID does not exist" Dec 04 13:13:18 crc kubenswrapper[4979]: I1204 13:13:18.810321 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-96578f89-dn7t8"] Dec 04 13:13:19 crc kubenswrapper[4979]: I1204 13:13:19.285174 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:19 crc kubenswrapper[4979]: I1204 13:13:19.433959 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5c694d4454-2bzkv" Dec 04 13:13:20 crc kubenswrapper[4979]: I1204 13:13:20.212458 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" path="/var/lib/kubelet/pods/7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee/volumes" Dec 04 13:13:28 crc kubenswrapper[4979]: I1204 13:13:28.040905 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:13:28 crc kubenswrapper[4979]: I1204 13:13:28.041522 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.881222 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-mbnxj"] Dec 04 13:13:30 crc kubenswrapper[4979]: E1204 13:13:30.882573 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerName="dnsmasq-dns" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.882691 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerName="dnsmasq-dns" Dec 04 13:13:30 crc kubenswrapper[4979]: E1204 13:13:30.882809 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerName="init" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.882886 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerName="init" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.883206 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ea5e9ee-d9bd-4395-86ff-ddbc52a9a3ee" containerName="dnsmasq-dns" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.884061 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.913198 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mbnxj"] Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.967792 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw5xl\" (UniqueName: \"kubernetes.io/projected/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-kube-api-access-jw5xl\") pod \"neutron-db-create-mbnxj\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.967905 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-operator-scripts\") pod \"neutron-db-create-mbnxj\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:30 crc kubenswrapper[4979]: I1204 13:13:30.998371 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-a773-account-create-update-t7h6h"] Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.000157 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.003897 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.046427 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a773-account-create-update-t7h6h"] Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.069462 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw5xl\" (UniqueName: \"kubernetes.io/projected/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-kube-api-access-jw5xl\") pod \"neutron-db-create-mbnxj\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.069546 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-operator-scripts\") pod \"neutron-db-create-mbnxj\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.070365 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-operator-scripts\") pod \"neutron-db-create-mbnxj\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.111633 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw5xl\" (UniqueName: \"kubernetes.io/projected/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-kube-api-access-jw5xl\") pod \"neutron-db-create-mbnxj\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.172371 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f887401-69a4-44e6-af00-97e2208c8137-operator-scripts\") pod \"neutron-a773-account-create-update-t7h6h\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.172474 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4srq\" (UniqueName: \"kubernetes.io/projected/7f887401-69a4-44e6-af00-97e2208c8137-kube-api-access-l4srq\") pod \"neutron-a773-account-create-update-t7h6h\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.210654 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.274449 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f887401-69a4-44e6-af00-97e2208c8137-operator-scripts\") pod \"neutron-a773-account-create-update-t7h6h\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.274526 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4srq\" (UniqueName: \"kubernetes.io/projected/7f887401-69a4-44e6-af00-97e2208c8137-kube-api-access-l4srq\") pod \"neutron-a773-account-create-update-t7h6h\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.275511 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f887401-69a4-44e6-af00-97e2208c8137-operator-scripts\") pod \"neutron-a773-account-create-update-t7h6h\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.294683 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4srq\" (UniqueName: \"kubernetes.io/projected/7f887401-69a4-44e6-af00-97e2208c8137-kube-api-access-l4srq\") pod \"neutron-a773-account-create-update-t7h6h\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.335525 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.672024 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-mbnxj"] Dec 04 13:13:31 crc kubenswrapper[4979]: W1204 13:13:31.680702 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb5673d2_1830_410f_a8e5_ff1ff68ccd4b.slice/crio-1a93ac578ad80da5af9a176d3e2d2ee6cd2f4e94bfadecdea3a3e3df522edebd WatchSource:0}: Error finding container 1a93ac578ad80da5af9a176d3e2d2ee6cd2f4e94bfadecdea3a3e3df522edebd: Status 404 returned error can't find the container with id 1a93ac578ad80da5af9a176d3e2d2ee6cd2f4e94bfadecdea3a3e3df522edebd Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.791464 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-a773-account-create-update-t7h6h"] Dec 04 13:13:31 crc kubenswrapper[4979]: W1204 13:13:31.797875 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f887401_69a4_44e6_af00_97e2208c8137.slice/crio-5b46734af9e1a01bde5952d90c2dfaa6f8cf11b7de26072a0c68549eec997095 WatchSource:0}: Error finding container 5b46734af9e1a01bde5952d90c2dfaa6f8cf11b7de26072a0c68549eec997095: Status 404 returned error can't find the container with id 5b46734af9e1a01bde5952d90c2dfaa6f8cf11b7de26072a0c68549eec997095 Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.858627 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a773-account-create-update-t7h6h" event={"ID":"7f887401-69a4-44e6-af00-97e2208c8137","Type":"ContainerStarted","Data":"5b46734af9e1a01bde5952d90c2dfaa6f8cf11b7de26072a0c68549eec997095"} Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.861269 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mbnxj" event={"ID":"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b","Type":"ContainerStarted","Data":"2d272ff8edba3842c49c6a817a8ab18970d87dc215ec67a73bb759a47cc5b909"} Dec 04 13:13:31 crc kubenswrapper[4979]: I1204 13:13:31.861322 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mbnxj" event={"ID":"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b","Type":"ContainerStarted","Data":"1a93ac578ad80da5af9a176d3e2d2ee6cd2f4e94bfadecdea3a3e3df522edebd"} Dec 04 13:13:32 crc kubenswrapper[4979]: I1204 13:13:32.870403 4979 generic.go:334] "Generic (PLEG): container finished" podID="bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" containerID="2d272ff8edba3842c49c6a817a8ab18970d87dc215ec67a73bb759a47cc5b909" exitCode=0 Dec 04 13:13:32 crc kubenswrapper[4979]: I1204 13:13:32.870448 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mbnxj" event={"ID":"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b","Type":"ContainerDied","Data":"2d272ff8edba3842c49c6a817a8ab18970d87dc215ec67a73bb759a47cc5b909"} Dec 04 13:13:32 crc kubenswrapper[4979]: I1204 13:13:32.871952 4979 generic.go:334] "Generic (PLEG): container finished" podID="7f887401-69a4-44e6-af00-97e2208c8137" containerID="0b7e96c223001574ccd4e65dc7936441d4f164b47fee37c67117d33eb723e09a" exitCode=0 Dec 04 13:13:32 crc kubenswrapper[4979]: I1204 13:13:32.872003 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a773-account-create-update-t7h6h" event={"ID":"7f887401-69a4-44e6-af00-97e2208c8137","Type":"ContainerDied","Data":"0b7e96c223001574ccd4e65dc7936441d4f164b47fee37c67117d33eb723e09a"} Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.299636 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.305550 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.431055 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-operator-scripts\") pod \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.431624 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f887401-69a4-44e6-af00-97e2208c8137-operator-scripts\") pod \"7f887401-69a4-44e6-af00-97e2208c8137\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.431712 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw5xl\" (UniqueName: \"kubernetes.io/projected/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-kube-api-access-jw5xl\") pod \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\" (UID: \"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b\") " Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.431823 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4srq\" (UniqueName: \"kubernetes.io/projected/7f887401-69a4-44e6-af00-97e2208c8137-kube-api-access-l4srq\") pod \"7f887401-69a4-44e6-af00-97e2208c8137\" (UID: \"7f887401-69a4-44e6-af00-97e2208c8137\") " Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.431954 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" (UID: "bb5673d2-1830-410f-a8e5-ff1ff68ccd4b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.432257 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f887401-69a4-44e6-af00-97e2208c8137-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7f887401-69a4-44e6-af00-97e2208c8137" (UID: "7f887401-69a4-44e6-af00-97e2208c8137"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.432573 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.432616 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7f887401-69a4-44e6-af00-97e2208c8137-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.438429 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f887401-69a4-44e6-af00-97e2208c8137-kube-api-access-l4srq" (OuterVolumeSpecName: "kube-api-access-l4srq") pod "7f887401-69a4-44e6-af00-97e2208c8137" (UID: "7f887401-69a4-44e6-af00-97e2208c8137"). InnerVolumeSpecName "kube-api-access-l4srq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.438508 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-kube-api-access-jw5xl" (OuterVolumeSpecName: "kube-api-access-jw5xl") pod "bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" (UID: "bb5673d2-1830-410f-a8e5-ff1ff68ccd4b"). InnerVolumeSpecName "kube-api-access-jw5xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.534725 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw5xl\" (UniqueName: \"kubernetes.io/projected/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b-kube-api-access-jw5xl\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.534774 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4srq\" (UniqueName: \"kubernetes.io/projected/7f887401-69a4-44e6-af00-97e2208c8137-kube-api-access-l4srq\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.889404 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-mbnxj" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.889420 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-mbnxj" event={"ID":"bb5673d2-1830-410f-a8e5-ff1ff68ccd4b","Type":"ContainerDied","Data":"1a93ac578ad80da5af9a176d3e2d2ee6cd2f4e94bfadecdea3a3e3df522edebd"} Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.889454 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a93ac578ad80da5af9a176d3e2d2ee6cd2f4e94bfadecdea3a3e3df522edebd" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.891729 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-a773-account-create-update-t7h6h" event={"ID":"7f887401-69a4-44e6-af00-97e2208c8137","Type":"ContainerDied","Data":"5b46734af9e1a01bde5952d90c2dfaa6f8cf11b7de26072a0c68549eec997095"} Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.891758 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b46734af9e1a01bde5952d90c2dfaa6f8cf11b7de26072a0c68549eec997095" Dec 04 13:13:34 crc kubenswrapper[4979]: I1204 13:13:34.891791 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-a773-account-create-update-t7h6h" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.223669 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-2crhq"] Dec 04 13:13:36 crc kubenswrapper[4979]: E1204 13:13:36.224088 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f887401-69a4-44e6-af00-97e2208c8137" containerName="mariadb-account-create-update" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.224107 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f887401-69a4-44e6-af00-97e2208c8137" containerName="mariadb-account-create-update" Dec 04 13:13:36 crc kubenswrapper[4979]: E1204 13:13:36.224162 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" containerName="mariadb-database-create" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.224172 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" containerName="mariadb-database-create" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.224376 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" containerName="mariadb-database-create" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.224403 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f887401-69a4-44e6-af00-97e2208c8137" containerName="mariadb-account-create-update" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.225001 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.229766 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.229906 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-m45qk" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.230066 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.231670 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2crhq"] Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.363557 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-combined-ca-bundle\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.364420 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2clgl\" (UniqueName: \"kubernetes.io/projected/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-kube-api-access-2clgl\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.364491 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-config\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.465652 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-combined-ca-bundle\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.465728 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2clgl\" (UniqueName: \"kubernetes.io/projected/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-kube-api-access-2clgl\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.465760 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-config\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.470937 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-config\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.471888 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-combined-ca-bundle\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.487958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2clgl\" (UniqueName: \"kubernetes.io/projected/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-kube-api-access-2clgl\") pod \"neutron-db-sync-2crhq\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:36 crc kubenswrapper[4979]: I1204 13:13:36.548096 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:37 crc kubenswrapper[4979]: I1204 13:13:37.012903 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-2crhq"] Dec 04 13:13:37 crc kubenswrapper[4979]: I1204 13:13:37.931184 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2crhq" event={"ID":"1174e2d1-da2d-4e5d-a1a0-9852a10457ee","Type":"ContainerStarted","Data":"76af69893ea3fc1b963f5f63db2c5133081472afa4514a0861b2318ba91ad508"} Dec 04 13:13:37 crc kubenswrapper[4979]: I1204 13:13:37.931627 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2crhq" event={"ID":"1174e2d1-da2d-4e5d-a1a0-9852a10457ee","Type":"ContainerStarted","Data":"403e67125c131e24c9c529d6e823afc0a19cf6139ff8f1aa0a23f32bb392bede"} Dec 04 13:13:37 crc kubenswrapper[4979]: I1204 13:13:37.952803 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-2crhq" podStartSLOduration=1.952777116 podStartE2EDuration="1.952777116s" podCreationTimestamp="2025-12-04 13:13:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:37.94767667 +0000 UTC m=+5442.221972484" watchObservedRunningTime="2025-12-04 13:13:37.952777116 +0000 UTC m=+5442.227072920" Dec 04 13:13:43 crc kubenswrapper[4979]: I1204 13:13:43.402109 4979 generic.go:334] "Generic (PLEG): container finished" podID="1174e2d1-da2d-4e5d-a1a0-9852a10457ee" containerID="76af69893ea3fc1b963f5f63db2c5133081472afa4514a0861b2318ba91ad508" exitCode=0 Dec 04 13:13:43 crc kubenswrapper[4979]: I1204 13:13:43.402201 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2crhq" event={"ID":"1174e2d1-da2d-4e5d-a1a0-9852a10457ee","Type":"ContainerDied","Data":"76af69893ea3fc1b963f5f63db2c5133081472afa4514a0861b2318ba91ad508"} Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.701458 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.822695 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-combined-ca-bundle\") pod \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.822775 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-config\") pod \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.822833 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2clgl\" (UniqueName: \"kubernetes.io/projected/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-kube-api-access-2clgl\") pod \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\" (UID: \"1174e2d1-da2d-4e5d-a1a0-9852a10457ee\") " Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.828552 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-kube-api-access-2clgl" (OuterVolumeSpecName: "kube-api-access-2clgl") pod "1174e2d1-da2d-4e5d-a1a0-9852a10457ee" (UID: "1174e2d1-da2d-4e5d-a1a0-9852a10457ee"). InnerVolumeSpecName "kube-api-access-2clgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.846174 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-config" (OuterVolumeSpecName: "config") pod "1174e2d1-da2d-4e5d-a1a0-9852a10457ee" (UID: "1174e2d1-da2d-4e5d-a1a0-9852a10457ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.847822 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1174e2d1-da2d-4e5d-a1a0-9852a10457ee" (UID: "1174e2d1-da2d-4e5d-a1a0-9852a10457ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.925886 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.925960 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:44 crc kubenswrapper[4979]: I1204 13:13:44.925974 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2clgl\" (UniqueName: \"kubernetes.io/projected/1174e2d1-da2d-4e5d-a1a0-9852a10457ee-kube-api-access-2clgl\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.420938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-2crhq" event={"ID":"1174e2d1-da2d-4e5d-a1a0-9852a10457ee","Type":"ContainerDied","Data":"403e67125c131e24c9c529d6e823afc0a19cf6139ff8f1aa0a23f32bb392bede"} Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.421362 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="403e67125c131e24c9c529d6e823afc0a19cf6139ff8f1aa0a23f32bb392bede" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.421201 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-2crhq" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.643724 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6889464645-4xlcb"] Dec 04 13:13:45 crc kubenswrapper[4979]: E1204 13:13:45.644097 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1174e2d1-da2d-4e5d-a1a0-9852a10457ee" containerName="neutron-db-sync" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.644116 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1174e2d1-da2d-4e5d-a1a0-9852a10457ee" containerName="neutron-db-sync" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.644355 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1174e2d1-da2d-4e5d-a1a0-9852a10457ee" containerName="neutron-db-sync" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.645853 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.655799 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6889464645-4xlcb"] Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.742810 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp8kh\" (UniqueName: \"kubernetes.io/projected/37992729-2f06-4d26-a32c-5535e21b9f2a-kube-api-access-bp8kh\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.742896 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-sb\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.742930 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-config\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.742960 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-dns-svc\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.743004 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-nb\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.763375 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5968f5544f-zdvnf"] Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.765279 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.774352 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.776526 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.776694 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-m45qk" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.784784 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5968f5544f-zdvnf"] Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.844627 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-sb\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.844696 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-config\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.844733 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-dns-svc\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.844779 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-nb\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.844946 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-config\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.844971 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-combined-ca-bundle\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.845006 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp8kh\" (UniqueName: \"kubernetes.io/projected/37992729-2f06-4d26-a32c-5535e21b9f2a-kube-api-access-bp8kh\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.845034 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fncjc\" (UniqueName: \"kubernetes.io/projected/3e7547b4-b2b6-4917-b872-1282d7a8a876-kube-api-access-fncjc\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.845057 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-httpd-config\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.846117 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-sb\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.846823 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-config\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.847486 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-dns-svc\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.848046 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-nb\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.881666 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp8kh\" (UniqueName: \"kubernetes.io/projected/37992729-2f06-4d26-a32c-5535e21b9f2a-kube-api-access-bp8kh\") pod \"dnsmasq-dns-6889464645-4xlcb\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.946622 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-config\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.946720 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-combined-ca-bundle\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.946764 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fncjc\" (UniqueName: \"kubernetes.io/projected/3e7547b4-b2b6-4917-b872-1282d7a8a876-kube-api-access-fncjc\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.946787 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-httpd-config\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.953388 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-config\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.958164 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-httpd-config\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.971607 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.971663 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e7547b4-b2b6-4917-b872-1282d7a8a876-combined-ca-bundle\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:45 crc kubenswrapper[4979]: I1204 13:13:45.973529 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fncjc\" (UniqueName: \"kubernetes.io/projected/3e7547b4-b2b6-4917-b872-1282d7a8a876-kube-api-access-fncjc\") pod \"neutron-5968f5544f-zdvnf\" (UID: \"3e7547b4-b2b6-4917-b872-1282d7a8a876\") " pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:46 crc kubenswrapper[4979]: I1204 13:13:46.128235 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:46 crc kubenswrapper[4979]: I1204 13:13:46.465874 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6889464645-4xlcb"] Dec 04 13:13:46 crc kubenswrapper[4979]: I1204 13:13:46.688271 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5968f5544f-zdvnf"] Dec 04 13:13:46 crc kubenswrapper[4979]: W1204 13:13:46.693706 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e7547b4_b2b6_4917_b872_1282d7a8a876.slice/crio-198c49cf3e43e6cb59d2893c717658db665aa5d67f19a0a2da93dc95e2183ddc WatchSource:0}: Error finding container 198c49cf3e43e6cb59d2893c717658db665aa5d67f19a0a2da93dc95e2183ddc: Status 404 returned error can't find the container with id 198c49cf3e43e6cb59d2893c717658db665aa5d67f19a0a2da93dc95e2183ddc Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.437703 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5968f5544f-zdvnf" event={"ID":"3e7547b4-b2b6-4917-b872-1282d7a8a876","Type":"ContainerStarted","Data":"9a136bf81d2d76c6833173623ff734e56ae2ec7f0823d39c7eb24df0c1828bf0"} Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.438314 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.438346 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5968f5544f-zdvnf" event={"ID":"3e7547b4-b2b6-4917-b872-1282d7a8a876","Type":"ContainerStarted","Data":"4266556e08dd3d24278b4a869d58b5e60cfa4b646f205097b70f4394275c1a30"} Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.438356 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5968f5544f-zdvnf" event={"ID":"3e7547b4-b2b6-4917-b872-1282d7a8a876","Type":"ContainerStarted","Data":"198c49cf3e43e6cb59d2893c717658db665aa5d67f19a0a2da93dc95e2183ddc"} Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.439415 4979 generic.go:334] "Generic (PLEG): container finished" podID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerID="4d632a4ea0d11cac821ffbbdd9cd12b2f919b44e517201d96e3e4db74e7ea60e" exitCode=0 Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.439446 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6889464645-4xlcb" event={"ID":"37992729-2f06-4d26-a32c-5535e21b9f2a","Type":"ContainerDied","Data":"4d632a4ea0d11cac821ffbbdd9cd12b2f919b44e517201d96e3e4db74e7ea60e"} Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.439464 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6889464645-4xlcb" event={"ID":"37992729-2f06-4d26-a32c-5535e21b9f2a","Type":"ContainerStarted","Data":"0f3259cfdee3c64c3d5c13403b9b0dd6f9558ffbe0bd3a8a6a96cff6c8e82aba"} Dec 04 13:13:47 crc kubenswrapper[4979]: I1204 13:13:47.483534 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5968f5544f-zdvnf" podStartSLOduration=2.483514198 podStartE2EDuration="2.483514198s" podCreationTimestamp="2025-12-04 13:13:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:47.461932354 +0000 UTC m=+5451.736228158" watchObservedRunningTime="2025-12-04 13:13:47.483514198 +0000 UTC m=+5451.757810002" Dec 04 13:13:48 crc kubenswrapper[4979]: I1204 13:13:48.449648 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6889464645-4xlcb" event={"ID":"37992729-2f06-4d26-a32c-5535e21b9f2a","Type":"ContainerStarted","Data":"4ed5d5cdd4cdb901449d565cd86a646df0e5515e208efbf79dab2e8cc7af1c2b"} Dec 04 13:13:48 crc kubenswrapper[4979]: I1204 13:13:48.484922 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6889464645-4xlcb" podStartSLOduration=3.484893748 podStartE2EDuration="3.484893748s" podCreationTimestamp="2025-12-04 13:13:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:13:48.472479158 +0000 UTC m=+5452.746774962" watchObservedRunningTime="2025-12-04 13:13:48.484893748 +0000 UTC m=+5452.759189552" Dec 04 13:13:49 crc kubenswrapper[4979]: I1204 13:13:49.464185 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:55 crc kubenswrapper[4979]: I1204 13:13:55.973529 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.035384 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8c6d95f7-bxwbv"] Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.035979 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerName="dnsmasq-dns" containerID="cri-o://7064714334f098d81b367567cebbb47b023b8e2ff638bdaf9537aa2d9dddfaa3" gracePeriod=10 Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.526964 4979 generic.go:334] "Generic (PLEG): container finished" podID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerID="7064714334f098d81b367567cebbb47b023b8e2ff638bdaf9537aa2d9dddfaa3" exitCode=0 Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.527022 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" event={"ID":"d7e6b5a2-0b10-449e-ae95-11dfebf8f956","Type":"ContainerDied","Data":"7064714334f098d81b367567cebbb47b023b8e2ff638bdaf9537aa2d9dddfaa3"} Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.527448 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" event={"ID":"d7e6b5a2-0b10-449e-ae95-11dfebf8f956","Type":"ContainerDied","Data":"91bdc0da5348a3b0b7366b876e7d76b651c0ac25a5adb1604b98ca3d2de92578"} Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.527474 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91bdc0da5348a3b0b7366b876e7d76b651c0ac25a5adb1604b98ca3d2de92578" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.565242 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.645766 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-nb\") pod \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.646110 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-config\") pod \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.646192 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-dns-svc\") pod \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.646345 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlvfd\" (UniqueName: \"kubernetes.io/projected/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-kube-api-access-tlvfd\") pod \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.646429 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-sb\") pod \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\" (UID: \"d7e6b5a2-0b10-449e-ae95-11dfebf8f956\") " Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.669732 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-kube-api-access-tlvfd" (OuterVolumeSpecName: "kube-api-access-tlvfd") pod "d7e6b5a2-0b10-449e-ae95-11dfebf8f956" (UID: "d7e6b5a2-0b10-449e-ae95-11dfebf8f956"). InnerVolumeSpecName "kube-api-access-tlvfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.689518 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-config" (OuterVolumeSpecName: "config") pod "d7e6b5a2-0b10-449e-ae95-11dfebf8f956" (UID: "d7e6b5a2-0b10-449e-ae95-11dfebf8f956"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.694940 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d7e6b5a2-0b10-449e-ae95-11dfebf8f956" (UID: "d7e6b5a2-0b10-449e-ae95-11dfebf8f956"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.701846 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d7e6b5a2-0b10-449e-ae95-11dfebf8f956" (UID: "d7e6b5a2-0b10-449e-ae95-11dfebf8f956"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.703747 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d7e6b5a2-0b10-449e-ae95-11dfebf8f956" (UID: "d7e6b5a2-0b10-449e-ae95-11dfebf8f956"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.749712 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.749743 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.749752 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.749761 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlvfd\" (UniqueName: \"kubernetes.io/projected/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-kube-api-access-tlvfd\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:56 crc kubenswrapper[4979]: I1204 13:13:56.749786 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d7e6b5a2-0b10-449e-ae95-11dfebf8f956-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:13:57 crc kubenswrapper[4979]: I1204 13:13:57.536722 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6d95f7-bxwbv" Dec 04 13:13:57 crc kubenswrapper[4979]: I1204 13:13:57.578717 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8c6d95f7-bxwbv"] Dec 04 13:13:57 crc kubenswrapper[4979]: I1204 13:13:57.586435 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8c6d95f7-bxwbv"] Dec 04 13:13:58 crc kubenswrapper[4979]: I1204 13:13:58.041089 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:13:58 crc kubenswrapper[4979]: I1204 13:13:58.041153 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:13:58 crc kubenswrapper[4979]: I1204 13:13:58.210532 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" path="/var/lib/kubelet/pods/d7e6b5a2-0b10-449e-ae95-11dfebf8f956/volumes" Dec 04 13:14:16 crc kubenswrapper[4979]: I1204 13:14:16.137590 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5968f5544f-zdvnf" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.244966 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-6j2nx"] Dec 04 13:14:23 crc kubenswrapper[4979]: E1204 13:14:23.246024 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerName="dnsmasq-dns" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.246043 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerName="dnsmasq-dns" Dec 04 13:14:23 crc kubenswrapper[4979]: E1204 13:14:23.246070 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerName="init" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.246078 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerName="init" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.246285 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7e6b5a2-0b10-449e-ae95-11dfebf8f956" containerName="dnsmasq-dns" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.247038 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.255041 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6j2nx"] Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.334240 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/074dff34-e984-42cf-a467-ac5e96716fec-operator-scripts\") pod \"glance-db-create-6j2nx\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.334383 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vzvm\" (UniqueName: \"kubernetes.io/projected/074dff34-e984-42cf-a467-ac5e96716fec-kube-api-access-7vzvm\") pod \"glance-db-create-6j2nx\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.338600 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-e2de-account-create-update-gmnz9"] Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.340229 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.344773 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.357913 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e2de-account-create-update-gmnz9"] Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.435733 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vzvm\" (UniqueName: \"kubernetes.io/projected/074dff34-e984-42cf-a467-ac5e96716fec-kube-api-access-7vzvm\") pod \"glance-db-create-6j2nx\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.435806 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lkrs\" (UniqueName: \"kubernetes.io/projected/540d2569-0399-40eb-a451-e43cf58270bc-kube-api-access-6lkrs\") pod \"glance-e2de-account-create-update-gmnz9\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.435926 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/074dff34-e984-42cf-a467-ac5e96716fec-operator-scripts\") pod \"glance-db-create-6j2nx\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.435962 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/540d2569-0399-40eb-a451-e43cf58270bc-operator-scripts\") pod \"glance-e2de-account-create-update-gmnz9\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.437250 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/074dff34-e984-42cf-a467-ac5e96716fec-operator-scripts\") pod \"glance-db-create-6j2nx\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.457967 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vzvm\" (UniqueName: \"kubernetes.io/projected/074dff34-e984-42cf-a467-ac5e96716fec-kube-api-access-7vzvm\") pod \"glance-db-create-6j2nx\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.537260 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/540d2569-0399-40eb-a451-e43cf58270bc-operator-scripts\") pod \"glance-e2de-account-create-update-gmnz9\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.537423 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lkrs\" (UniqueName: \"kubernetes.io/projected/540d2569-0399-40eb-a451-e43cf58270bc-kube-api-access-6lkrs\") pod \"glance-e2de-account-create-update-gmnz9\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.538153 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/540d2569-0399-40eb-a451-e43cf58270bc-operator-scripts\") pod \"glance-e2de-account-create-update-gmnz9\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.554052 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lkrs\" (UniqueName: \"kubernetes.io/projected/540d2569-0399-40eb-a451-e43cf58270bc-kube-api-access-6lkrs\") pod \"glance-e2de-account-create-update-gmnz9\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.568009 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:23 crc kubenswrapper[4979]: I1204 13:14:23.658971 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.010245 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-6j2nx"] Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.116083 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-e2de-account-create-update-gmnz9"] Dec 04 13:14:24 crc kubenswrapper[4979]: W1204 13:14:24.122482 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod540d2569_0399_40eb_a451_e43cf58270bc.slice/crio-f6634b83ff53e5ef62a3d68a0ba34e6a64fc0cad2365aef28b50d14ef42e09a7 WatchSource:0}: Error finding container f6634b83ff53e5ef62a3d68a0ba34e6a64fc0cad2365aef28b50d14ef42e09a7: Status 404 returned error can't find the container with id f6634b83ff53e5ef62a3d68a0ba34e6a64fc0cad2365aef28b50d14ef42e09a7 Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.770230 4979 generic.go:334] "Generic (PLEG): container finished" podID="074dff34-e984-42cf-a467-ac5e96716fec" containerID="28c2582ba35345340d3fbe83f285b843289af55b4163fcc7e49865be02c4a8ee" exitCode=0 Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.770367 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6j2nx" event={"ID":"074dff34-e984-42cf-a467-ac5e96716fec","Type":"ContainerDied","Data":"28c2582ba35345340d3fbe83f285b843289af55b4163fcc7e49865be02c4a8ee"} Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.771396 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6j2nx" event={"ID":"074dff34-e984-42cf-a467-ac5e96716fec","Type":"ContainerStarted","Data":"f518704fe3ffcfbb5d636984946ab8b4a7265b88b312cceb7f2cca5faf475315"} Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.773798 4979 generic.go:334] "Generic (PLEG): container finished" podID="540d2569-0399-40eb-a451-e43cf58270bc" containerID="fa195f06e4e43bbbbb877669aa5457baf431b7f3a779e0f1cc1d35e2d85c6388" exitCode=0 Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.773833 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e2de-account-create-update-gmnz9" event={"ID":"540d2569-0399-40eb-a451-e43cf58270bc","Type":"ContainerDied","Data":"fa195f06e4e43bbbbb877669aa5457baf431b7f3a779e0f1cc1d35e2d85c6388"} Dec 04 13:14:24 crc kubenswrapper[4979]: I1204 13:14:24.773865 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e2de-account-create-update-gmnz9" event={"ID":"540d2569-0399-40eb-a451-e43cf58270bc","Type":"ContainerStarted","Data":"f6634b83ff53e5ef62a3d68a0ba34e6a64fc0cad2365aef28b50d14ef42e09a7"} Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.135679 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.144678 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.187326 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vzvm\" (UniqueName: \"kubernetes.io/projected/074dff34-e984-42cf-a467-ac5e96716fec-kube-api-access-7vzvm\") pod \"074dff34-e984-42cf-a467-ac5e96716fec\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.187780 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/540d2569-0399-40eb-a451-e43cf58270bc-operator-scripts\") pod \"540d2569-0399-40eb-a451-e43cf58270bc\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.187853 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lkrs\" (UniqueName: \"kubernetes.io/projected/540d2569-0399-40eb-a451-e43cf58270bc-kube-api-access-6lkrs\") pod \"540d2569-0399-40eb-a451-e43cf58270bc\" (UID: \"540d2569-0399-40eb-a451-e43cf58270bc\") " Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.187904 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/074dff34-e984-42cf-a467-ac5e96716fec-operator-scripts\") pod \"074dff34-e984-42cf-a467-ac5e96716fec\" (UID: \"074dff34-e984-42cf-a467-ac5e96716fec\") " Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.188915 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/540d2569-0399-40eb-a451-e43cf58270bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "540d2569-0399-40eb-a451-e43cf58270bc" (UID: "540d2569-0399-40eb-a451-e43cf58270bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.190123 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/074dff34-e984-42cf-a467-ac5e96716fec-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "074dff34-e984-42cf-a467-ac5e96716fec" (UID: "074dff34-e984-42cf-a467-ac5e96716fec"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.194231 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/074dff34-e984-42cf-a467-ac5e96716fec-kube-api-access-7vzvm" (OuterVolumeSpecName: "kube-api-access-7vzvm") pod "074dff34-e984-42cf-a467-ac5e96716fec" (UID: "074dff34-e984-42cf-a467-ac5e96716fec"). InnerVolumeSpecName "kube-api-access-7vzvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.194386 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540d2569-0399-40eb-a451-e43cf58270bc-kube-api-access-6lkrs" (OuterVolumeSpecName: "kube-api-access-6lkrs") pod "540d2569-0399-40eb-a451-e43cf58270bc" (UID: "540d2569-0399-40eb-a451-e43cf58270bc"). InnerVolumeSpecName "kube-api-access-6lkrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.291193 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vzvm\" (UniqueName: \"kubernetes.io/projected/074dff34-e984-42cf-a467-ac5e96716fec-kube-api-access-7vzvm\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.291342 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/540d2569-0399-40eb-a451-e43cf58270bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.291356 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lkrs\" (UniqueName: \"kubernetes.io/projected/540d2569-0399-40eb-a451-e43cf58270bc-kube-api-access-6lkrs\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.291368 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/074dff34-e984-42cf-a467-ac5e96716fec-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.792321 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-e2de-account-create-update-gmnz9" event={"ID":"540d2569-0399-40eb-a451-e43cf58270bc","Type":"ContainerDied","Data":"f6634b83ff53e5ef62a3d68a0ba34e6a64fc0cad2365aef28b50d14ef42e09a7"} Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.792367 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6634b83ff53e5ef62a3d68a0ba34e6a64fc0cad2365aef28b50d14ef42e09a7" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.792725 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-e2de-account-create-update-gmnz9" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.793963 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-6j2nx" event={"ID":"074dff34-e984-42cf-a467-ac5e96716fec","Type":"ContainerDied","Data":"f518704fe3ffcfbb5d636984946ab8b4a7265b88b312cceb7f2cca5faf475315"} Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.793984 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f518704fe3ffcfbb5d636984946ab8b4a7265b88b312cceb7f2cca5faf475315" Dec 04 13:14:26 crc kubenswrapper[4979]: I1204 13:14:26.794029 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-6j2nx" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.040726 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.041033 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.041094 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.041919 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f41cc400d512fbc84293fd7f01e9e77498ba3dd4183c76c64ed639782564147d"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.041995 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://f41cc400d512fbc84293fd7f01e9e77498ba3dd4183c76c64ed639782564147d" gracePeriod=600 Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.539737 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-tcn4f"] Dec 04 13:14:28 crc kubenswrapper[4979]: E1204 13:14:28.540347 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="074dff34-e984-42cf-a467-ac5e96716fec" containerName="mariadb-database-create" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.540360 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="074dff34-e984-42cf-a467-ac5e96716fec" containerName="mariadb-database-create" Dec 04 13:14:28 crc kubenswrapper[4979]: E1204 13:14:28.540373 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="540d2569-0399-40eb-a451-e43cf58270bc" containerName="mariadb-account-create-update" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.540379 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="540d2569-0399-40eb-a451-e43cf58270bc" containerName="mariadb-account-create-update" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.540567 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="074dff34-e984-42cf-a467-ac5e96716fec" containerName="mariadb-database-create" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.540595 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="540d2569-0399-40eb-a451-e43cf58270bc" containerName="mariadb-account-create-update" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.541147 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.543682 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gbs9f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.545611 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.553731 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tcn4f"] Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.625863 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-config-data\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.625920 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9rl7\" (UniqueName: \"kubernetes.io/projected/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-kube-api-access-z9rl7\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.625985 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-db-sync-config-data\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.626018 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-combined-ca-bundle\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.727848 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-config-data\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.729320 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9rl7\" (UniqueName: \"kubernetes.io/projected/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-kube-api-access-z9rl7\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.729431 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-db-sync-config-data\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.729485 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-combined-ca-bundle\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.734680 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-db-sync-config-data\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.738350 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-config-data\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.747947 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-combined-ca-bundle\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.753103 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9rl7\" (UniqueName: \"kubernetes.io/projected/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-kube-api-access-z9rl7\") pod \"glance-db-sync-tcn4f\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.811676 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="f41cc400d512fbc84293fd7f01e9e77498ba3dd4183c76c64ed639782564147d" exitCode=0 Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.811724 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"f41cc400d512fbc84293fd7f01e9e77498ba3dd4183c76c64ed639782564147d"} Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.811763 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005"} Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.811783 4979 scope.go:117] "RemoveContainer" containerID="e529f8682a5fe7616136c86fa7b4ab25966bf13abd5c311fa28b8caee74623cc" Dec 04 13:14:28 crc kubenswrapper[4979]: I1204 13:14:28.871157 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:29 crc kubenswrapper[4979]: I1204 13:14:29.390252 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-tcn4f"] Dec 04 13:14:29 crc kubenswrapper[4979]: I1204 13:14:29.825531 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tcn4f" event={"ID":"0522ef0b-76e7-4c30-af05-d7f3f4eb3154","Type":"ContainerStarted","Data":"dbc5a0a6305a24e9283146d16571291d8048e9ead54e19ee66acca8a7eb3688b"} Dec 04 13:14:30 crc kubenswrapper[4979]: I1204 13:14:30.839167 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tcn4f" event={"ID":"0522ef0b-76e7-4c30-af05-d7f3f4eb3154","Type":"ContainerStarted","Data":"a82cef0dea9e1aa128f51638af2ab6923edf92bdea5fbfb84c6fd95c6e0d4ae0"} Dec 04 13:14:30 crc kubenswrapper[4979]: I1204 13:14:30.858953 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-tcn4f" podStartSLOduration=2.858930876 podStartE2EDuration="2.858930876s" podCreationTimestamp="2025-12-04 13:14:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:14:30.857815136 +0000 UTC m=+5495.132110940" watchObservedRunningTime="2025-12-04 13:14:30.858930876 +0000 UTC m=+5495.133226680" Dec 04 13:14:33 crc kubenswrapper[4979]: I1204 13:14:33.867371 4979 generic.go:334] "Generic (PLEG): container finished" podID="0522ef0b-76e7-4c30-af05-d7f3f4eb3154" containerID="a82cef0dea9e1aa128f51638af2ab6923edf92bdea5fbfb84c6fd95c6e0d4ae0" exitCode=0 Dec 04 13:14:33 crc kubenswrapper[4979]: I1204 13:14:33.867590 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tcn4f" event={"ID":"0522ef0b-76e7-4c30-af05-d7f3f4eb3154","Type":"ContainerDied","Data":"a82cef0dea9e1aa128f51638af2ab6923edf92bdea5fbfb84c6fd95c6e0d4ae0"} Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.247279 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.338603 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-combined-ca-bundle\") pod \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.338727 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9rl7\" (UniqueName: \"kubernetes.io/projected/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-kube-api-access-z9rl7\") pod \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.338834 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-config-data\") pod \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.338889 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-db-sync-config-data\") pod \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\" (UID: \"0522ef0b-76e7-4c30-af05-d7f3f4eb3154\") " Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.346603 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-kube-api-access-z9rl7" (OuterVolumeSpecName: "kube-api-access-z9rl7") pod "0522ef0b-76e7-4c30-af05-d7f3f4eb3154" (UID: "0522ef0b-76e7-4c30-af05-d7f3f4eb3154"). InnerVolumeSpecName "kube-api-access-z9rl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.347243 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0522ef0b-76e7-4c30-af05-d7f3f4eb3154" (UID: "0522ef0b-76e7-4c30-af05-d7f3f4eb3154"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.365247 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0522ef0b-76e7-4c30-af05-d7f3f4eb3154" (UID: "0522ef0b-76e7-4c30-af05-d7f3f4eb3154"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.384457 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-config-data" (OuterVolumeSpecName: "config-data") pod "0522ef0b-76e7-4c30-af05-d7f3f4eb3154" (UID: "0522ef0b-76e7-4c30-af05-d7f3f4eb3154"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.443046 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.443646 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9rl7\" (UniqueName: \"kubernetes.io/projected/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-kube-api-access-z9rl7\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.443755 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.443852 4979 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0522ef0b-76e7-4c30-af05-d7f3f4eb3154-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.886804 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-tcn4f" event={"ID":"0522ef0b-76e7-4c30-af05-d7f3f4eb3154","Type":"ContainerDied","Data":"dbc5a0a6305a24e9283146d16571291d8048e9ead54e19ee66acca8a7eb3688b"} Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.886854 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbc5a0a6305a24e9283146d16571291d8048e9ead54e19ee66acca8a7eb3688b" Dec 04 13:14:35 crc kubenswrapper[4979]: I1204 13:14:35.886868 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-tcn4f" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.267173 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b79f8c4df-sc469"] Dec 04 13:14:36 crc kubenswrapper[4979]: E1204 13:14:36.267739 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0522ef0b-76e7-4c30-af05-d7f3f4eb3154" containerName="glance-db-sync" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.267757 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0522ef0b-76e7-4c30-af05-d7f3f4eb3154" containerName="glance-db-sync" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.268345 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0522ef0b-76e7-4c30-af05-d7f3f4eb3154" containerName="glance-db-sync" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.280200 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.282975 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-sb\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.283103 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q59jj\" (UniqueName: \"kubernetes.io/projected/089bfb86-563e-4fea-88b7-845414efa9fe-kube-api-access-q59jj\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.283232 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-dns-svc\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.283286 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-nb\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.283327 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-config\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.301858 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.306071 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.315933 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.316109 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.316177 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gbs9f" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.316284 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.317449 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b79f8c4df-sc469"] Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.334228 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.383279 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384664 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384707 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-dns-svc\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384746 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384775 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddd2g\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-kube-api-access-ddd2g\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384795 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-nb\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384815 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-config\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384874 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-sb\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384931 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q59jj\" (UniqueName: \"kubernetes.io/projected/089bfb86-563e-4fea-88b7-845414efa9fe-kube-api-access-q59jj\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384942 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384958 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-ceph\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.384990 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-logs\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.385008 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.385038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.385643 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-dns-svc\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.385828 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-nb\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.386386 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-sb\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.388684 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-config\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.389562 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.390077 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.412355 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q59jj\" (UniqueName: \"kubernetes.io/projected/089bfb86-563e-4fea-88b7-845414efa9fe-kube-api-access-q59jj\") pod \"dnsmasq-dns-5b79f8c4df-sc469\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487460 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-logs\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487712 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487741 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487776 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487821 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487855 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487875 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487898 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddd2g\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-kube-api-access-ddd2g\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487943 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.487976 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkcd2\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-kube-api-access-dkcd2\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.488001 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.488042 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-ceph\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.488061 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-ceph\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.488092 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.488531 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-logs\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.488861 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.492924 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-ceph\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.494916 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-config-data\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.495587 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.497446 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-scripts\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.512124 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddd2g\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-kube-api-access-ddd2g\") pod \"glance-default-external-api-0\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.589699 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.589862 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.589912 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkcd2\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-kube-api-access-dkcd2\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.589937 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.589954 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-ceph\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.589985 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.590001 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.591679 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-logs\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.592443 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.595436 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-ceph\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.595766 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.596127 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.596363 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.613875 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkcd2\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-kube-api-access-dkcd2\") pod \"glance-default-internal-api-0\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.634242 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.654239 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:14:36 crc kubenswrapper[4979]: I1204 13:14:36.752288 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.154703 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b79f8c4df-sc469"] Dec 04 13:14:37 crc kubenswrapper[4979]: W1204 13:14:37.175533 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30a8e385_329f_4a6c_9384_cd5aeb4f17e6.slice/crio-1d63ccdac15ba2cee15a1675bd19a913baf23581bf93571d4df7f161266ee0ee WatchSource:0}: Error finding container 1d63ccdac15ba2cee15a1675bd19a913baf23581bf93571d4df7f161266ee0ee: Status 404 returned error can't find the container with id 1d63ccdac15ba2cee15a1675bd19a913baf23581bf93571d4df7f161266ee0ee Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.177190 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.303222 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.440986 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:37 crc kubenswrapper[4979]: W1204 13:14:37.502545 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2574eaf3_801c_4c6d_8f65_15f908b643e6.slice/crio-21254d06f435d0c8662542c9a07a1cb841c5c2d9122a89767a82df79e68e99e1 WatchSource:0}: Error finding container 21254d06f435d0c8662542c9a07a1cb841c5c2d9122a89767a82df79e68e99e1: Status 404 returned error can't find the container with id 21254d06f435d0c8662542c9a07a1cb841c5c2d9122a89767a82df79e68e99e1 Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.916829 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"30a8e385-329f-4a6c-9384-cd5aeb4f17e6","Type":"ContainerStarted","Data":"1d63ccdac15ba2cee15a1675bd19a913baf23581bf93571d4df7f161266ee0ee"} Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.918323 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2574eaf3-801c-4c6d-8f65-15f908b643e6","Type":"ContainerStarted","Data":"21254d06f435d0c8662542c9a07a1cb841c5c2d9122a89767a82df79e68e99e1"} Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.921148 4979 generic.go:334] "Generic (PLEG): container finished" podID="089bfb86-563e-4fea-88b7-845414efa9fe" containerID="fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf" exitCode=0 Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.921196 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" event={"ID":"089bfb86-563e-4fea-88b7-845414efa9fe","Type":"ContainerDied","Data":"fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf"} Dec 04 13:14:37 crc kubenswrapper[4979]: I1204 13:14:37.921223 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" event={"ID":"089bfb86-563e-4fea-88b7-845414efa9fe","Type":"ContainerStarted","Data":"38d25de7bad7b7510eb9353c9c03cfd9a83f885c11594b23d60436f83edbc433"} Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.930997 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2574eaf3-801c-4c6d-8f65-15f908b643e6","Type":"ContainerStarted","Data":"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009"} Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.931570 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2574eaf3-801c-4c6d-8f65-15f908b643e6","Type":"ContainerStarted","Data":"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517"} Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.933072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" event={"ID":"089bfb86-563e-4fea-88b7-845414efa9fe","Type":"ContainerStarted","Data":"c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885"} Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.933344 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.935816 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"30a8e385-329f-4a6c-9384-cd5aeb4f17e6","Type":"ContainerStarted","Data":"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab"} Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.935860 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"30a8e385-329f-4a6c-9384-cd5aeb4f17e6","Type":"ContainerStarted","Data":"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa"} Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.936111 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-httpd" containerID="cri-o://8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab" gracePeriod=30 Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.936092 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-log" containerID="cri-o://58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa" gracePeriod=30 Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.959956 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=2.959931451 podStartE2EDuration="2.959931451s" podCreationTimestamp="2025-12-04 13:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:14:38.956126859 +0000 UTC m=+5503.230422663" watchObservedRunningTime="2025-12-04 13:14:38.959931451 +0000 UTC m=+5503.234227255" Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.979346 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=2.979318706 podStartE2EDuration="2.979318706s" podCreationTimestamp="2025-12-04 13:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:14:38.97493815 +0000 UTC m=+5503.249233974" watchObservedRunningTime="2025-12-04 13:14:38.979318706 +0000 UTC m=+5503.253614510" Dec 04 13:14:38 crc kubenswrapper[4979]: I1204 13:14:38.998397 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" podStartSLOduration=2.998376972 podStartE2EDuration="2.998376972s" podCreationTimestamp="2025-12-04 13:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:14:38.99789104 +0000 UTC m=+5503.272186854" watchObservedRunningTime="2025-12-04 13:14:38.998376972 +0000 UTC m=+5503.272672776" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.588870 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.665065 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.781684 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-logs\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.781739 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-combined-ca-bundle\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.781779 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddd2g\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-kube-api-access-ddd2g\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.781889 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-config-data\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.782295 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-logs" (OuterVolumeSpecName: "logs") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.782688 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-scripts\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.782769 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-httpd-run\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.782866 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-ceph\") pod \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\" (UID: \"30a8e385-329f-4a6c-9384-cd5aeb4f17e6\") " Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.783418 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.783803 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.788620 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-ceph" (OuterVolumeSpecName: "ceph") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.789264 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-scripts" (OuterVolumeSpecName: "scripts") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.789630 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-kube-api-access-ddd2g" (OuterVolumeSpecName: "kube-api-access-ddd2g") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "kube-api-access-ddd2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.809963 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.833592 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-config-data" (OuterVolumeSpecName: "config-data") pod "30a8e385-329f-4a6c-9384-cd5aeb4f17e6" (UID: "30a8e385-329f-4a6c-9384-cd5aeb4f17e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.885946 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.886096 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.886175 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.886259 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.886357 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.886458 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddd2g\" (UniqueName: \"kubernetes.io/projected/30a8e385-329f-4a6c-9384-cd5aeb4f17e6-kube-api-access-ddd2g\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.947284 4979 generic.go:334] "Generic (PLEG): container finished" podID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerID="8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab" exitCode=0 Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.947338 4979 generic.go:334] "Generic (PLEG): container finished" podID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerID="58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa" exitCode=143 Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.947407 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"30a8e385-329f-4a6c-9384-cd5aeb4f17e6","Type":"ContainerDied","Data":"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab"} Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.947481 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"30a8e385-329f-4a6c-9384-cd5aeb4f17e6","Type":"ContainerDied","Data":"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa"} Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.947497 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"30a8e385-329f-4a6c-9384-cd5aeb4f17e6","Type":"ContainerDied","Data":"1d63ccdac15ba2cee15a1675bd19a913baf23581bf93571d4df7f161266ee0ee"} Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.947519 4979 scope.go:117] "RemoveContainer" containerID="8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.948526 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:14:39 crc kubenswrapper[4979]: I1204 13:14:39.974367 4979 scope.go:117] "RemoveContainer" containerID="58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:39.996269 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.008748 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.012573 4979 scope.go:117] "RemoveContainer" containerID="8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab" Dec 04 13:14:40 crc kubenswrapper[4979]: E1204 13:14:40.013279 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab\": container with ID starting with 8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab not found: ID does not exist" containerID="8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.013347 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab"} err="failed to get container status \"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab\": rpc error: code = NotFound desc = could not find container \"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab\": container with ID starting with 8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab not found: ID does not exist" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.013377 4979 scope.go:117] "RemoveContainer" containerID="58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa" Dec 04 13:14:40 crc kubenswrapper[4979]: E1204 13:14:40.014040 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa\": container with ID starting with 58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa not found: ID does not exist" containerID="58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.014067 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa"} err="failed to get container status \"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa\": rpc error: code = NotFound desc = could not find container \"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa\": container with ID starting with 58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa not found: ID does not exist" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.014110 4979 scope.go:117] "RemoveContainer" containerID="8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.014714 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab"} err="failed to get container status \"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab\": rpc error: code = NotFound desc = could not find container \"8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab\": container with ID starting with 8935530bb53ffea95220054d9b2b22e8df11c0f1eea0a1eaad4762b4b4b756ab not found: ID does not exist" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.014731 4979 scope.go:117] "RemoveContainer" containerID="58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.014949 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa"} err="failed to get container status \"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa\": rpc error: code = NotFound desc = could not find container \"58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa\": container with ID starting with 58d5ef178841994bc7a52db0a170d381af84d4599b564adfae208d65c64feefa not found: ID does not exist" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.019948 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:40 crc kubenswrapper[4979]: E1204 13:14:40.020712 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-httpd" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.020743 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-httpd" Dec 04 13:14:40 crc kubenswrapper[4979]: E1204 13:14:40.020793 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-log" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.020806 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-log" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.021017 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-log" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.021058 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" containerName="glance-httpd" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.023754 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.028579 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.054065 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.090921 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.090993 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.091047 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.091097 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.091119 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vspjb\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-kube-api-access-vspjb\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.091134 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.091166 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-logs\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192411 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192451 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192509 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192577 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192598 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vspjb\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-kube-api-access-vspjb\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192615 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.192644 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-logs\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.193080 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.193096 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-logs\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.197259 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-ceph\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.197714 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.199766 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-scripts\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.200552 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-config-data\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.211496 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30a8e385-329f-4a6c-9384-cd5aeb4f17e6" path="/var/lib/kubelet/pods/30a8e385-329f-4a6c-9384-cd5aeb4f17e6/volumes" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.214677 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vspjb\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-kube-api-access-vspjb\") pod \"glance-default-external-api-0\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.349025 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.897569 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:14:40 crc kubenswrapper[4979]: W1204 13:14:40.905077 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod513a2a30_d753_457f_b030_3a40a2a2f8eb.slice/crio-64b3d98276797fdbb752831b12c050c7fc0d8bb7373d0056739e7ec29fabace2 WatchSource:0}: Error finding container 64b3d98276797fdbb752831b12c050c7fc0d8bb7373d0056739e7ec29fabace2: Status 404 returned error can't find the container with id 64b3d98276797fdbb752831b12c050c7fc0d8bb7373d0056739e7ec29fabace2 Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.961841 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"513a2a30-d753-457f-b030-3a40a2a2f8eb","Type":"ContainerStarted","Data":"64b3d98276797fdbb752831b12c050c7fc0d8bb7373d0056739e7ec29fabace2"} Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.962082 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-log" containerID="cri-o://dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517" gracePeriod=30 Dec 04 13:14:40 crc kubenswrapper[4979]: I1204 13:14:40.962157 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-httpd" containerID="cri-o://b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009" gracePeriod=30 Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.487708 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516521 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-logs\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516607 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-ceph\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516670 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-httpd-run\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516725 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-scripts\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516763 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-combined-ca-bundle\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516800 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkcd2\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-kube-api-access-dkcd2\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.516833 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-config-data\") pod \"2574eaf3-801c-4c6d-8f65-15f908b643e6\" (UID: \"2574eaf3-801c-4c6d-8f65-15f908b643e6\") " Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.518446 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-logs" (OuterVolumeSpecName: "logs") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.519279 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.529164 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-ceph" (OuterVolumeSpecName: "ceph") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.542461 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-scripts" (OuterVolumeSpecName: "scripts") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.554658 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-kube-api-access-dkcd2" (OuterVolumeSpecName: "kube-api-access-dkcd2") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "kube-api-access-dkcd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.568125 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.596761 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-config-data" (OuterVolumeSpecName: "config-data") pod "2574eaf3-801c-4c6d-8f65-15f908b643e6" (UID: "2574eaf3-801c-4c6d-8f65-15f908b643e6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618036 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618072 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618084 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618095 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2574eaf3-801c-4c6d-8f65-15f908b643e6-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618106 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618117 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2574eaf3-801c-4c6d-8f65-15f908b643e6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.618131 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkcd2\" (UniqueName: \"kubernetes.io/projected/2574eaf3-801c-4c6d-8f65-15f908b643e6-kube-api-access-dkcd2\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977720 4979 generic.go:334] "Generic (PLEG): container finished" podID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerID="b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009" exitCode=0 Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977776 4979 generic.go:334] "Generic (PLEG): container finished" podID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerID="dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517" exitCode=143 Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977846 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2574eaf3-801c-4c6d-8f65-15f908b643e6","Type":"ContainerDied","Data":"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009"} Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977941 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2574eaf3-801c-4c6d-8f65-15f908b643e6","Type":"ContainerDied","Data":"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517"} Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977961 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2574eaf3-801c-4c6d-8f65-15f908b643e6","Type":"ContainerDied","Data":"21254d06f435d0c8662542c9a07a1cb841c5c2d9122a89767a82df79e68e99e1"} Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977982 4979 scope.go:117] "RemoveContainer" containerID="b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.977870 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:41 crc kubenswrapper[4979]: I1204 13:14:41.982018 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"513a2a30-d753-457f-b030-3a40a2a2f8eb","Type":"ContainerStarted","Data":"c247c615c9dfd23d374587ba506083ce2c138d5972ecdca3fbcc960f05b4f49b"} Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.019111 4979 scope.go:117] "RemoveContainer" containerID="dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.030794 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.046335 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.049943 4979 scope.go:117] "RemoveContainer" containerID="b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009" Dec 04 13:14:42 crc kubenswrapper[4979]: E1204 13:14:42.050610 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009\": container with ID starting with b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009 not found: ID does not exist" containerID="b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.050647 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009"} err="failed to get container status \"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009\": rpc error: code = NotFound desc = could not find container \"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009\": container with ID starting with b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009 not found: ID does not exist" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.050676 4979 scope.go:117] "RemoveContainer" containerID="dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517" Dec 04 13:14:42 crc kubenswrapper[4979]: E1204 13:14:42.051040 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517\": container with ID starting with dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517 not found: ID does not exist" containerID="dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.051077 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517"} err="failed to get container status \"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517\": rpc error: code = NotFound desc = could not find container \"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517\": container with ID starting with dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517 not found: ID does not exist" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.051098 4979 scope.go:117] "RemoveContainer" containerID="b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.052492 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009"} err="failed to get container status \"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009\": rpc error: code = NotFound desc = could not find container \"b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009\": container with ID starting with b0016a8ab22720ab9c9bd8d8f0e4774bec643046802f6c85b1f10aff6e61b009 not found: ID does not exist" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.052627 4979 scope.go:117] "RemoveContainer" containerID="dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.053077 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517"} err="failed to get container status \"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517\": rpc error: code = NotFound desc = could not find container \"dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517\": container with ID starting with dd046c02ad37d31f50c9e0962c18cbe871ee940f0cb39d58074ae5ee21d21517 not found: ID does not exist" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.056469 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:42 crc kubenswrapper[4979]: E1204 13:14:42.057866 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-log" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.057904 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-log" Dec 04 13:14:42 crc kubenswrapper[4979]: E1204 13:14:42.057956 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-httpd" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.057966 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-httpd" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.058175 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-httpd" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.058230 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" containerName="glance-log" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.059450 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.062260 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.066600 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.133041 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.133691 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.133888 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-logs\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.134180 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.134324 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkgts\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-kube-api-access-lkgts\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.134458 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.134620 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.210191 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2574eaf3-801c-4c6d-8f65-15f908b643e6" path="/var/lib/kubelet/pods/2574eaf3-801c-4c6d-8f65-15f908b643e6/volumes" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.236778 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-logs\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.236949 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkgts\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-kube-api-access-lkgts\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.236982 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.237013 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.237046 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.237096 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.237171 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.237471 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-logs\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.239045 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.242128 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-ceph\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.242158 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.243080 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.244374 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.280093 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkgts\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-kube-api-access-lkgts\") pod \"glance-default-internal-api-0\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.387624 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:42 crc kubenswrapper[4979]: I1204 13:14:42.996029 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"513a2a30-d753-457f-b030-3a40a2a2f8eb","Type":"ContainerStarted","Data":"73d277822e69dc649a1304a039dab984e18070400cece29139dde97ae693b2d8"} Dec 04 13:14:43 crc kubenswrapper[4979]: I1204 13:14:43.024094 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.024069931 podStartE2EDuration="4.024069931s" podCreationTimestamp="2025-12-04 13:14:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:14:43.016991013 +0000 UTC m=+5507.291286817" watchObservedRunningTime="2025-12-04 13:14:43.024069931 +0000 UTC m=+5507.298365735" Dec 04 13:14:43 crc kubenswrapper[4979]: I1204 13:14:43.089717 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:14:44 crc kubenswrapper[4979]: I1204 13:14:44.014965 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6817c34-ecd6-4e01-8968-ad815733d564","Type":"ContainerStarted","Data":"2397952bdf020e37361aff8cdeedb643db349cada1efde5eb5f2216a061cf75d"} Dec 04 13:14:44 crc kubenswrapper[4979]: I1204 13:14:44.015608 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6817c34-ecd6-4e01-8968-ad815733d564","Type":"ContainerStarted","Data":"401e783b59f75fe92a9e431f4aea337f82d9a83e23dc30cf8e4b39ac26109cf2"} Dec 04 13:14:45 crc kubenswrapper[4979]: I1204 13:14:45.025694 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6817c34-ecd6-4e01-8968-ad815733d564","Type":"ContainerStarted","Data":"f801e69ff572d809c6a02b485a314f214f9bcbfcbc2a040df5c53900140cae16"} Dec 04 13:14:45 crc kubenswrapper[4979]: I1204 13:14:45.058808 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.058783221 podStartE2EDuration="3.058783221s" podCreationTimestamp="2025-12-04 13:14:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:14:45.044719017 +0000 UTC m=+5509.319014861" watchObservedRunningTime="2025-12-04 13:14:45.058783221 +0000 UTC m=+5509.333079035" Dec 04 13:14:46 crc kubenswrapper[4979]: I1204 13:14:46.637486 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:14:46 crc kubenswrapper[4979]: I1204 13:14:46.719704 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6889464645-4xlcb"] Dec 04 13:14:46 crc kubenswrapper[4979]: I1204 13:14:46.725391 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6889464645-4xlcb" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerName="dnsmasq-dns" containerID="cri-o://4ed5d5cdd4cdb901449d565cd86a646df0e5515e208efbf79dab2e8cc7af1c2b" gracePeriod=10 Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.043367 4979 generic.go:334] "Generic (PLEG): container finished" podID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerID="4ed5d5cdd4cdb901449d565cd86a646df0e5515e208efbf79dab2e8cc7af1c2b" exitCode=0 Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.043422 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6889464645-4xlcb" event={"ID":"37992729-2f06-4d26-a32c-5535e21b9f2a","Type":"ContainerDied","Data":"4ed5d5cdd4cdb901449d565cd86a646df0e5515e208efbf79dab2e8cc7af1c2b"} Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.700498 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.844819 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-sb\") pod \"37992729-2f06-4d26-a32c-5535e21b9f2a\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.844866 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-nb\") pod \"37992729-2f06-4d26-a32c-5535e21b9f2a\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.844951 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-config\") pod \"37992729-2f06-4d26-a32c-5535e21b9f2a\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.845057 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp8kh\" (UniqueName: \"kubernetes.io/projected/37992729-2f06-4d26-a32c-5535e21b9f2a-kube-api-access-bp8kh\") pod \"37992729-2f06-4d26-a32c-5535e21b9f2a\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.845096 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-dns-svc\") pod \"37992729-2f06-4d26-a32c-5535e21b9f2a\" (UID: \"37992729-2f06-4d26-a32c-5535e21b9f2a\") " Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.851581 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37992729-2f06-4d26-a32c-5535e21b9f2a-kube-api-access-bp8kh" (OuterVolumeSpecName: "kube-api-access-bp8kh") pod "37992729-2f06-4d26-a32c-5535e21b9f2a" (UID: "37992729-2f06-4d26-a32c-5535e21b9f2a"). InnerVolumeSpecName "kube-api-access-bp8kh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.891583 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "37992729-2f06-4d26-a32c-5535e21b9f2a" (UID: "37992729-2f06-4d26-a32c-5535e21b9f2a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.892327 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "37992729-2f06-4d26-a32c-5535e21b9f2a" (UID: "37992729-2f06-4d26-a32c-5535e21b9f2a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.892919 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "37992729-2f06-4d26-a32c-5535e21b9f2a" (UID: "37992729-2f06-4d26-a32c-5535e21b9f2a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.906677 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-config" (OuterVolumeSpecName: "config") pod "37992729-2f06-4d26-a32c-5535e21b9f2a" (UID: "37992729-2f06-4d26-a32c-5535e21b9f2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.947116 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.947162 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.947175 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.947186 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37992729-2f06-4d26-a32c-5535e21b9f2a-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:47 crc kubenswrapper[4979]: I1204 13:14:47.947198 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp8kh\" (UniqueName: \"kubernetes.io/projected/37992729-2f06-4d26-a32c-5535e21b9f2a-kube-api-access-bp8kh\") on node \"crc\" DevicePath \"\"" Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.053030 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6889464645-4xlcb" event={"ID":"37992729-2f06-4d26-a32c-5535e21b9f2a","Type":"ContainerDied","Data":"0f3259cfdee3c64c3d5c13403b9b0dd6f9558ffbe0bd3a8a6a96cff6c8e82aba"} Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.053126 4979 scope.go:117] "RemoveContainer" containerID="4ed5d5cdd4cdb901449d565cd86a646df0e5515e208efbf79dab2e8cc7af1c2b" Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.053122 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6889464645-4xlcb" Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.075447 4979 scope.go:117] "RemoveContainer" containerID="4d632a4ea0d11cac821ffbbdd9cd12b2f919b44e517201d96e3e4db74e7ea60e" Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.086125 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6889464645-4xlcb"] Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.093362 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6889464645-4xlcb"] Dec 04 13:14:48 crc kubenswrapper[4979]: I1204 13:14:48.211457 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" path="/var/lib/kubelet/pods/37992729-2f06-4d26-a32c-5535e21b9f2a/volumes" Dec 04 13:14:48 crc kubenswrapper[4979]: E1204 13:14:48.279864 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod37992729_2f06_4d26_a32c_5535e21b9f2a.slice\": RecentStats: unable to find data in memory cache]" Dec 04 13:14:50 crc kubenswrapper[4979]: I1204 13:14:50.349240 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 13:14:50 crc kubenswrapper[4979]: I1204 13:14:50.349732 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 13:14:50 crc kubenswrapper[4979]: I1204 13:14:50.380290 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 13:14:50 crc kubenswrapper[4979]: I1204 13:14:50.396125 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 13:14:51 crc kubenswrapper[4979]: I1204 13:14:51.082263 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 13:14:51 crc kubenswrapper[4979]: I1204 13:14:51.082721 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 13:14:52 crc kubenswrapper[4979]: I1204 13:14:52.388085 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:52 crc kubenswrapper[4979]: I1204 13:14:52.388189 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:52 crc kubenswrapper[4979]: I1204 13:14:52.418146 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:52 crc kubenswrapper[4979]: I1204 13:14:52.431478 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:53 crc kubenswrapper[4979]: I1204 13:14:53.098175 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:53 crc kubenswrapper[4979]: I1204 13:14:53.098608 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:53 crc kubenswrapper[4979]: I1204 13:14:53.360059 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 13:14:53 crc kubenswrapper[4979]: I1204 13:14:53.360185 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:14:53 crc kubenswrapper[4979]: I1204 13:14:53.433483 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 13:14:55 crc kubenswrapper[4979]: I1204 13:14:55.325243 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 13:14:55 crc kubenswrapper[4979]: I1204 13:14:55.325753 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:14:55 crc kubenswrapper[4979]: I1204 13:14:55.327203 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.143751 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw"] Dec 04 13:15:00 crc kubenswrapper[4979]: E1204 13:15:00.144742 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerName="dnsmasq-dns" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.144758 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerName="dnsmasq-dns" Dec 04 13:15:00 crc kubenswrapper[4979]: E1204 13:15:00.144776 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerName="init" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.144782 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerName="init" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.144951 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="37992729-2f06-4d26-a32c-5535e21b9f2a" containerName="dnsmasq-dns" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.145633 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.148281 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.148517 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.155424 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw"] Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.281579 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2705356-a364-400a-95e3-8649b6a3e9aa-config-volume\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.281657 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdr2p\" (UniqueName: \"kubernetes.io/projected/b2705356-a364-400a-95e3-8649b6a3e9aa-kube-api-access-wdr2p\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.281677 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2705356-a364-400a-95e3-8649b6a3e9aa-secret-volume\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.383185 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2705356-a364-400a-95e3-8649b6a3e9aa-config-volume\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.383568 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdr2p\" (UniqueName: \"kubernetes.io/projected/b2705356-a364-400a-95e3-8649b6a3e9aa-kube-api-access-wdr2p\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.383686 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2705356-a364-400a-95e3-8649b6a3e9aa-secret-volume\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.384022 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2705356-a364-400a-95e3-8649b6a3e9aa-config-volume\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.389664 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2705356-a364-400a-95e3-8649b6a3e9aa-secret-volume\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.403931 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdr2p\" (UniqueName: \"kubernetes.io/projected/b2705356-a364-400a-95e3-8649b6a3e9aa-kube-api-access-wdr2p\") pod \"collect-profiles-29414235-782dw\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.464583 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:00 crc kubenswrapper[4979]: I1204 13:15:00.947048 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw"] Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.163667 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" event={"ID":"b2705356-a364-400a-95e3-8649b6a3e9aa","Type":"ContainerStarted","Data":"a4bdd8d43e3133278015fa8732f4379dcae9419af3c7f999589b416a4de2f9cc"} Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.659406 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-7kpc6"] Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.660638 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.678692 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7kpc6"] Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.751060 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-2d60-account-create-update-5vtf9"] Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.752531 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.755624 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.762067 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2d60-account-create-update-5vtf9"] Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.809757 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knd4b\" (UniqueName: \"kubernetes.io/projected/e7043a66-6d3c-423a-a159-84afd283492e-kube-api-access-knd4b\") pod \"placement-db-create-7kpc6\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.809956 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7043a66-6d3c-423a-a159-84afd283492e-operator-scripts\") pod \"placement-db-create-7kpc6\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.911232 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f58092-8c55-4172-b2b5-da79f085277c-operator-scripts\") pod \"placement-2d60-account-create-update-5vtf9\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.911618 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7043a66-6d3c-423a-a159-84afd283492e-operator-scripts\") pod \"placement-db-create-7kpc6\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.911730 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knd4b\" (UniqueName: \"kubernetes.io/projected/e7043a66-6d3c-423a-a159-84afd283492e-kube-api-access-knd4b\") pod \"placement-db-create-7kpc6\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.911805 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbft5\" (UniqueName: \"kubernetes.io/projected/a6f58092-8c55-4172-b2b5-da79f085277c-kube-api-access-pbft5\") pod \"placement-2d60-account-create-update-5vtf9\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.912585 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7043a66-6d3c-423a-a159-84afd283492e-operator-scripts\") pod \"placement-db-create-7kpc6\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.929817 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knd4b\" (UniqueName: \"kubernetes.io/projected/e7043a66-6d3c-423a-a159-84afd283492e-kube-api-access-knd4b\") pod \"placement-db-create-7kpc6\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:01 crc kubenswrapper[4979]: I1204 13:15:01.979932 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.012943 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbft5\" (UniqueName: \"kubernetes.io/projected/a6f58092-8c55-4172-b2b5-da79f085277c-kube-api-access-pbft5\") pod \"placement-2d60-account-create-update-5vtf9\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.012998 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f58092-8c55-4172-b2b5-da79f085277c-operator-scripts\") pod \"placement-2d60-account-create-update-5vtf9\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.013708 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f58092-8c55-4172-b2b5-da79f085277c-operator-scripts\") pod \"placement-2d60-account-create-update-5vtf9\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.031998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbft5\" (UniqueName: \"kubernetes.io/projected/a6f58092-8c55-4172-b2b5-da79f085277c-kube-api-access-pbft5\") pod \"placement-2d60-account-create-update-5vtf9\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.070000 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.178072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" event={"ID":"b2705356-a364-400a-95e3-8649b6a3e9aa","Type":"ContainerStarted","Data":"6fb6c72a6ff784c768c49fa34ee00ca6c2acfe2df20d84c7c6cc6833efd83b03"} Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.438114 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" podStartSLOduration=2.438090976 podStartE2EDuration="2.438090976s" podCreationTimestamp="2025-12-04 13:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:15:02.207744173 +0000 UTC m=+5526.482039977" watchObservedRunningTime="2025-12-04 13:15:02.438090976 +0000 UTC m=+5526.712386780" Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.441563 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-7kpc6"] Dec 04 13:15:02 crc kubenswrapper[4979]: I1204 13:15:02.572989 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-2d60-account-create-update-5vtf9"] Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.188701 4979 generic.go:334] "Generic (PLEG): container finished" podID="b2705356-a364-400a-95e3-8649b6a3e9aa" containerID="6fb6c72a6ff784c768c49fa34ee00ca6c2acfe2df20d84c7c6cc6833efd83b03" exitCode=0 Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.188817 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" event={"ID":"b2705356-a364-400a-95e3-8649b6a3e9aa","Type":"ContainerDied","Data":"6fb6c72a6ff784c768c49fa34ee00ca6c2acfe2df20d84c7c6cc6833efd83b03"} Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.191288 4979 generic.go:334] "Generic (PLEG): container finished" podID="e7043a66-6d3c-423a-a159-84afd283492e" containerID="2eba55fc70887aaf607b256f7a7f1a7939db931f489376c840e3549523186743" exitCode=0 Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.191386 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7kpc6" event={"ID":"e7043a66-6d3c-423a-a159-84afd283492e","Type":"ContainerDied","Data":"2eba55fc70887aaf607b256f7a7f1a7939db931f489376c840e3549523186743"} Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.191415 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7kpc6" event={"ID":"e7043a66-6d3c-423a-a159-84afd283492e","Type":"ContainerStarted","Data":"2decc83f4b49fe55f3c244c89f4339f70e956fc39430e130409fe9e2b4849852"} Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.192935 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6f58092-8c55-4172-b2b5-da79f085277c" containerID="af27934eeef91f7c26f220e51ad6e7a0708a3f9d6c356153b2cf2eb76c747a1e" exitCode=0 Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.192979 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d60-account-create-update-5vtf9" event={"ID":"a6f58092-8c55-4172-b2b5-da79f085277c","Type":"ContainerDied","Data":"af27934eeef91f7c26f220e51ad6e7a0708a3f9d6c356153b2cf2eb76c747a1e"} Dec 04 13:15:03 crc kubenswrapper[4979]: I1204 13:15:03.193002 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d60-account-create-update-5vtf9" event={"ID":"a6f58092-8c55-4172-b2b5-da79f085277c","Type":"ContainerStarted","Data":"38a2cd46a5d21f248774b9666698aea1abba66bb5337ce7d512a50c48949dac4"} Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.614222 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.623676 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.633634 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.761866 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbft5\" (UniqueName: \"kubernetes.io/projected/a6f58092-8c55-4172-b2b5-da79f085277c-kube-api-access-pbft5\") pod \"a6f58092-8c55-4172-b2b5-da79f085277c\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.762356 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f58092-8c55-4172-b2b5-da79f085277c-operator-scripts\") pod \"a6f58092-8c55-4172-b2b5-da79f085277c\" (UID: \"a6f58092-8c55-4172-b2b5-da79f085277c\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.762600 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knd4b\" (UniqueName: \"kubernetes.io/projected/e7043a66-6d3c-423a-a159-84afd283492e-kube-api-access-knd4b\") pod \"e7043a66-6d3c-423a-a159-84afd283492e\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.762715 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2705356-a364-400a-95e3-8649b6a3e9aa-secret-volume\") pod \"b2705356-a364-400a-95e3-8649b6a3e9aa\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.762873 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7043a66-6d3c-423a-a159-84afd283492e-operator-scripts\") pod \"e7043a66-6d3c-423a-a159-84afd283492e\" (UID: \"e7043a66-6d3c-423a-a159-84afd283492e\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.763024 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2705356-a364-400a-95e3-8649b6a3e9aa-config-volume\") pod \"b2705356-a364-400a-95e3-8649b6a3e9aa\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.762944 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6f58092-8c55-4172-b2b5-da79f085277c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a6f58092-8c55-4172-b2b5-da79f085277c" (UID: "a6f58092-8c55-4172-b2b5-da79f085277c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.763373 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7043a66-6d3c-423a-a159-84afd283492e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e7043a66-6d3c-423a-a159-84afd283492e" (UID: "e7043a66-6d3c-423a-a159-84afd283492e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.763565 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2705356-a364-400a-95e3-8649b6a3e9aa-config-volume" (OuterVolumeSpecName: "config-volume") pod "b2705356-a364-400a-95e3-8649b6a3e9aa" (UID: "b2705356-a364-400a-95e3-8649b6a3e9aa"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.763779 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdr2p\" (UniqueName: \"kubernetes.io/projected/b2705356-a364-400a-95e3-8649b6a3e9aa-kube-api-access-wdr2p\") pod \"b2705356-a364-400a-95e3-8649b6a3e9aa\" (UID: \"b2705356-a364-400a-95e3-8649b6a3e9aa\") " Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.764550 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a6f58092-8c55-4172-b2b5-da79f085277c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.764595 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7043a66-6d3c-423a-a159-84afd283492e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.764614 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b2705356-a364-400a-95e3-8649b6a3e9aa-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.767980 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2705356-a364-400a-95e3-8649b6a3e9aa-kube-api-access-wdr2p" (OuterVolumeSpecName: "kube-api-access-wdr2p") pod "b2705356-a364-400a-95e3-8649b6a3e9aa" (UID: "b2705356-a364-400a-95e3-8649b6a3e9aa"). InnerVolumeSpecName "kube-api-access-wdr2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.768192 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6f58092-8c55-4172-b2b5-da79f085277c-kube-api-access-pbft5" (OuterVolumeSpecName: "kube-api-access-pbft5") pod "a6f58092-8c55-4172-b2b5-da79f085277c" (UID: "a6f58092-8c55-4172-b2b5-da79f085277c"). InnerVolumeSpecName "kube-api-access-pbft5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.768319 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7043a66-6d3c-423a-a159-84afd283492e-kube-api-access-knd4b" (OuterVolumeSpecName: "kube-api-access-knd4b") pod "e7043a66-6d3c-423a-a159-84afd283492e" (UID: "e7043a66-6d3c-423a-a159-84afd283492e"). InnerVolumeSpecName "kube-api-access-knd4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.768318 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b2705356-a364-400a-95e3-8649b6a3e9aa-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b2705356-a364-400a-95e3-8649b6a3e9aa" (UID: "b2705356-a364-400a-95e3-8649b6a3e9aa"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.865955 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbft5\" (UniqueName: \"kubernetes.io/projected/a6f58092-8c55-4172-b2b5-da79f085277c-kube-api-access-pbft5\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.865989 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knd4b\" (UniqueName: \"kubernetes.io/projected/e7043a66-6d3c-423a-a159-84afd283492e-kube-api-access-knd4b\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.866000 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b2705356-a364-400a-95e3-8649b6a3e9aa-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:04 crc kubenswrapper[4979]: I1204 13:15:04.866010 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdr2p\" (UniqueName: \"kubernetes.io/projected/b2705356-a364-400a-95e3-8649b6a3e9aa-kube-api-access-wdr2p\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.216010 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.216009 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw" event={"ID":"b2705356-a364-400a-95e3-8649b6a3e9aa","Type":"ContainerDied","Data":"a4bdd8d43e3133278015fa8732f4379dcae9419af3c7f999589b416a4de2f9cc"} Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.216590 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4bdd8d43e3133278015fa8732f4379dcae9419af3c7f999589b416a4de2f9cc" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.218483 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-2d60-account-create-update-5vtf9" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.218483 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-2d60-account-create-update-5vtf9" event={"ID":"a6f58092-8c55-4172-b2b5-da79f085277c","Type":"ContainerDied","Data":"38a2cd46a5d21f248774b9666698aea1abba66bb5337ce7d512a50c48949dac4"} Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.218521 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38a2cd46a5d21f248774b9666698aea1abba66bb5337ce7d512a50c48949dac4" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.220426 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-7kpc6" event={"ID":"e7043a66-6d3c-423a-a159-84afd283492e","Type":"ContainerDied","Data":"2decc83f4b49fe55f3c244c89f4339f70e956fc39430e130409fe9e2b4849852"} Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.220450 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2decc83f4b49fe55f3c244c89f4339f70e956fc39430e130409fe9e2b4849852" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.220474 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-7kpc6" Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.290910 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg"] Dec 04 13:15:05 crc kubenswrapper[4979]: I1204 13:15:05.298575 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414190-mqhlg"] Dec 04 13:15:06 crc kubenswrapper[4979]: I1204 13:15:06.208238 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a6d345d-320a-4625-88dc-17507072c856" path="/var/lib/kubelet/pods/4a6d345d-320a-4625-88dc-17507072c856/volumes" Dec 04 13:15:06 crc kubenswrapper[4979]: I1204 13:15:06.662578 4979 scope.go:117] "RemoveContainer" containerID="65274224dc2010c1de99ebd5b6d6eb2b84e48e3829a3444d156c4d22b5397cca" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.136528 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-f627f"] Dec 04 13:15:07 crc kubenswrapper[4979]: E1204 13:15:07.136901 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6f58092-8c55-4172-b2b5-da79f085277c" containerName="mariadb-account-create-update" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.136925 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6f58092-8c55-4172-b2b5-da79f085277c" containerName="mariadb-account-create-update" Dec 04 13:15:07 crc kubenswrapper[4979]: E1204 13:15:07.136946 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7043a66-6d3c-423a-a159-84afd283492e" containerName="mariadb-database-create" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.136952 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7043a66-6d3c-423a-a159-84afd283492e" containerName="mariadb-database-create" Dec 04 13:15:07 crc kubenswrapper[4979]: E1204 13:15:07.136962 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2705356-a364-400a-95e3-8649b6a3e9aa" containerName="collect-profiles" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.136968 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2705356-a364-400a-95e3-8649b6a3e9aa" containerName="collect-profiles" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.137131 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2705356-a364-400a-95e3-8649b6a3e9aa" containerName="collect-profiles" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.137152 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6f58092-8c55-4172-b2b5-da79f085277c" containerName="mariadb-account-create-update" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.137171 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7043a66-6d3c-423a-a159-84afd283492e" containerName="mariadb-database-create" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.137762 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.141269 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.141526 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-q7cz4" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.141643 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.165600 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-f627f"] Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.186538 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9db9c4bf-l6bqz"] Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.188403 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.203173 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-config-data\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.203220 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8c7039b-417d-4634-b1ca-8a145614d908-logs\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.203266 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9np4l\" (UniqueName: \"kubernetes.io/projected/d8c7039b-417d-4634-b1ca-8a145614d908-kube-api-access-9np4l\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.203359 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-scripts\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.203400 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-combined-ca-bundle\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.232556 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9db9c4bf-l6bqz"] Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305242 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-nb\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305379 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-combined-ca-bundle\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305465 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-config\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305489 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mggj\" (UniqueName: \"kubernetes.io/projected/1174334f-0570-4363-b95b-3870ca0ace66-kube-api-access-4mggj\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305518 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-config-data\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305557 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8c7039b-417d-4634-b1ca-8a145614d908-logs\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305629 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9np4l\" (UniqueName: \"kubernetes.io/projected/d8c7039b-417d-4634-b1ca-8a145614d908-kube-api-access-9np4l\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305672 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-dns-svc\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305696 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-sb\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.305799 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-scripts\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.306466 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8c7039b-417d-4634-b1ca-8a145614d908-logs\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.310505 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-scripts\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.311442 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-combined-ca-bundle\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.317253 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-config-data\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.328685 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9np4l\" (UniqueName: \"kubernetes.io/projected/d8c7039b-417d-4634-b1ca-8a145614d908-kube-api-access-9np4l\") pod \"placement-db-sync-f627f\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.407211 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-config\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.407257 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mggj\" (UniqueName: \"kubernetes.io/projected/1174334f-0570-4363-b95b-3870ca0ace66-kube-api-access-4mggj\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.407347 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-sb\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.407372 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-dns-svc\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.407458 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-nb\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.408471 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-sb\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.408493 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-nb\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.408604 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-dns-svc\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.408900 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-config\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.426257 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mggj\" (UniqueName: \"kubernetes.io/projected/1174334f-0570-4363-b95b-3870ca0ace66-kube-api-access-4mggj\") pod \"dnsmasq-dns-9db9c4bf-l6bqz\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.457390 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f627f" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.514239 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:07 crc kubenswrapper[4979]: I1204 13:15:07.900909 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-f627f"] Dec 04 13:15:08 crc kubenswrapper[4979]: I1204 13:15:08.018779 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9db9c4bf-l6bqz"] Dec 04 13:15:08 crc kubenswrapper[4979]: I1204 13:15:08.250207 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" event={"ID":"1174334f-0570-4363-b95b-3870ca0ace66","Type":"ContainerStarted","Data":"a0ba8a4ab6bf6dcc03d00d61a2040cf2298e367758166088c47363085515bb89"} Dec 04 13:15:08 crc kubenswrapper[4979]: I1204 13:15:08.251519 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f627f" event={"ID":"d8c7039b-417d-4634-b1ca-8a145614d908","Type":"ContainerStarted","Data":"fd95c52414337c1a25dac76c57b06176c67d2a9903868ce8a22ddcfcec7f9a4b"} Dec 04 13:15:08 crc kubenswrapper[4979]: I1204 13:15:08.251561 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f627f" event={"ID":"d8c7039b-417d-4634-b1ca-8a145614d908","Type":"ContainerStarted","Data":"925b07fb7401ea8a651d2e2191eb7417567784b8ec126640fb7ade14c2676db6"} Dec 04 13:15:08 crc kubenswrapper[4979]: I1204 13:15:08.279672 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-f627f" podStartSLOduration=1.279647606 podStartE2EDuration="1.279647606s" podCreationTimestamp="2025-12-04 13:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:15:08.270025031 +0000 UTC m=+5532.544320835" watchObservedRunningTime="2025-12-04 13:15:08.279647606 +0000 UTC m=+5532.553943410" Dec 04 13:15:09 crc kubenswrapper[4979]: I1204 13:15:09.261028 4979 generic.go:334] "Generic (PLEG): container finished" podID="1174334f-0570-4363-b95b-3870ca0ace66" containerID="3a6ed20a59ba4903324ea0804508a1173ae9d76988d18e4229ee3848b5d98f07" exitCode=0 Dec 04 13:15:09 crc kubenswrapper[4979]: I1204 13:15:09.261102 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" event={"ID":"1174334f-0570-4363-b95b-3870ca0ace66","Type":"ContainerDied","Data":"3a6ed20a59ba4903324ea0804508a1173ae9d76988d18e4229ee3848b5d98f07"} Dec 04 13:15:10 crc kubenswrapper[4979]: I1204 13:15:10.275010 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" event={"ID":"1174334f-0570-4363-b95b-3870ca0ace66","Type":"ContainerStarted","Data":"dd5acd3226af73f6fddcaea2f639962db5478838e225ca36a89084474d352db4"} Dec 04 13:15:10 crc kubenswrapper[4979]: I1204 13:15:10.276572 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:10 crc kubenswrapper[4979]: I1204 13:15:10.277676 4979 generic.go:334] "Generic (PLEG): container finished" podID="d8c7039b-417d-4634-b1ca-8a145614d908" containerID="fd95c52414337c1a25dac76c57b06176c67d2a9903868ce8a22ddcfcec7f9a4b" exitCode=0 Dec 04 13:15:10 crc kubenswrapper[4979]: I1204 13:15:10.277712 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f627f" event={"ID":"d8c7039b-417d-4634-b1ca-8a145614d908","Type":"ContainerDied","Data":"fd95c52414337c1a25dac76c57b06176c67d2a9903868ce8a22ddcfcec7f9a4b"} Dec 04 13:15:10 crc kubenswrapper[4979]: I1204 13:15:10.303351 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" podStartSLOduration=3.303330484 podStartE2EDuration="3.303330484s" podCreationTimestamp="2025-12-04 13:15:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:15:10.294252772 +0000 UTC m=+5534.568548576" watchObservedRunningTime="2025-12-04 13:15:10.303330484 +0000 UTC m=+5534.577626278" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.688542 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f627f" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.800766 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8c7039b-417d-4634-b1ca-8a145614d908-logs\") pod \"d8c7039b-417d-4634-b1ca-8a145614d908\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.800883 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-config-data\") pod \"d8c7039b-417d-4634-b1ca-8a145614d908\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.800937 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9np4l\" (UniqueName: \"kubernetes.io/projected/d8c7039b-417d-4634-b1ca-8a145614d908-kube-api-access-9np4l\") pod \"d8c7039b-417d-4634-b1ca-8a145614d908\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.801019 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-scripts\") pod \"d8c7039b-417d-4634-b1ca-8a145614d908\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.801069 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-combined-ca-bundle\") pod \"d8c7039b-417d-4634-b1ca-8a145614d908\" (UID: \"d8c7039b-417d-4634-b1ca-8a145614d908\") " Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.801833 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8c7039b-417d-4634-b1ca-8a145614d908-logs" (OuterVolumeSpecName: "logs") pod "d8c7039b-417d-4634-b1ca-8a145614d908" (UID: "d8c7039b-417d-4634-b1ca-8a145614d908"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.802110 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8c7039b-417d-4634-b1ca-8a145614d908-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.805888 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-scripts" (OuterVolumeSpecName: "scripts") pod "d8c7039b-417d-4634-b1ca-8a145614d908" (UID: "d8c7039b-417d-4634-b1ca-8a145614d908"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.806148 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8c7039b-417d-4634-b1ca-8a145614d908-kube-api-access-9np4l" (OuterVolumeSpecName: "kube-api-access-9np4l") pod "d8c7039b-417d-4634-b1ca-8a145614d908" (UID: "d8c7039b-417d-4634-b1ca-8a145614d908"). InnerVolumeSpecName "kube-api-access-9np4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.829541 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-config-data" (OuterVolumeSpecName: "config-data") pod "d8c7039b-417d-4634-b1ca-8a145614d908" (UID: "d8c7039b-417d-4634-b1ca-8a145614d908"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.831443 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8c7039b-417d-4634-b1ca-8a145614d908" (UID: "d8c7039b-417d-4634-b1ca-8a145614d908"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.904088 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.904126 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9np4l\" (UniqueName: \"kubernetes.io/projected/d8c7039b-417d-4634-b1ca-8a145614d908-kube-api-access-9np4l\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.904136 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:11 crc kubenswrapper[4979]: I1204 13:15:11.904147 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8c7039b-417d-4634-b1ca-8a145614d908-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.303105 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-f627f" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.303142 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-f627f" event={"ID":"d8c7039b-417d-4634-b1ca-8a145614d908","Type":"ContainerDied","Data":"925b07fb7401ea8a651d2e2191eb7417567784b8ec126640fb7ade14c2676db6"} Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.303233 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="925b07fb7401ea8a651d2e2191eb7417567784b8ec126640fb7ade14c2676db6" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.406604 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-69d694d89b-b86c2"] Dec 04 13:15:12 crc kubenswrapper[4979]: E1204 13:15:12.407079 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8c7039b-417d-4634-b1ca-8a145614d908" containerName="placement-db-sync" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.407092 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8c7039b-417d-4634-b1ca-8a145614d908" containerName="placement-db-sync" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.407324 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8c7039b-417d-4634-b1ca-8a145614d908" containerName="placement-db-sync" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.408357 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.411510 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-q7cz4" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.411606 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.411667 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.430675 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69d694d89b-b86c2"] Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.534158 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-config-data\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.534423 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cwmf\" (UniqueName: \"kubernetes.io/projected/833d82eb-d525-4813-8ad4-6ba2209832af-kube-api-access-9cwmf\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.534541 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-combined-ca-bundle\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.534629 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/833d82eb-d525-4813-8ad4-6ba2209832af-logs\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.534821 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-scripts\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.636004 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-combined-ca-bundle\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.636496 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/833d82eb-d525-4813-8ad4-6ba2209832af-logs\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.636581 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-scripts\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.636618 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-config-data\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.636677 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cwmf\" (UniqueName: \"kubernetes.io/projected/833d82eb-d525-4813-8ad4-6ba2209832af-kube-api-access-9cwmf\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.637415 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/833d82eb-d525-4813-8ad4-6ba2209832af-logs\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.642619 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-scripts\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.643249 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-combined-ca-bundle\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.643660 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/833d82eb-d525-4813-8ad4-6ba2209832af-config-data\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.655856 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cwmf\" (UniqueName: \"kubernetes.io/projected/833d82eb-d525-4813-8ad4-6ba2209832af-kube-api-access-9cwmf\") pod \"placement-69d694d89b-b86c2\" (UID: \"833d82eb-d525-4813-8ad4-6ba2209832af\") " pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:12 crc kubenswrapper[4979]: I1204 13:15:12.730770 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:13 crc kubenswrapper[4979]: I1204 13:15:13.239374 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-69d694d89b-b86c2"] Dec 04 13:15:13 crc kubenswrapper[4979]: W1204 13:15:13.246729 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod833d82eb_d525_4813_8ad4_6ba2209832af.slice/crio-818de2004fe5e8e363fc1078a93fd90b7deaf96a6aece7db48c30f08000557c8 WatchSource:0}: Error finding container 818de2004fe5e8e363fc1078a93fd90b7deaf96a6aece7db48c30f08000557c8: Status 404 returned error can't find the container with id 818de2004fe5e8e363fc1078a93fd90b7deaf96a6aece7db48c30f08000557c8 Dec 04 13:15:13 crc kubenswrapper[4979]: I1204 13:15:13.320554 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69d694d89b-b86c2" event={"ID":"833d82eb-d525-4813-8ad4-6ba2209832af","Type":"ContainerStarted","Data":"818de2004fe5e8e363fc1078a93fd90b7deaf96a6aece7db48c30f08000557c8"} Dec 04 13:15:14 crc kubenswrapper[4979]: I1204 13:15:14.332372 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69d694d89b-b86c2" event={"ID":"833d82eb-d525-4813-8ad4-6ba2209832af","Type":"ContainerStarted","Data":"1d046309794567f10fcfd2e36b9efe22c1c28b8d4c2f0f5c4c3a2b95d929bc2c"} Dec 04 13:15:14 crc kubenswrapper[4979]: I1204 13:15:14.332708 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-69d694d89b-b86c2" event={"ID":"833d82eb-d525-4813-8ad4-6ba2209832af","Type":"ContainerStarted","Data":"a894236aac1d371a56c42226f25558ac2316f39d8b84f13c732c55491be9cc24"} Dec 04 13:15:14 crc kubenswrapper[4979]: I1204 13:15:14.332948 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:14 crc kubenswrapper[4979]: I1204 13:15:14.373998 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-69d694d89b-b86c2" podStartSLOduration=2.373977176 podStartE2EDuration="2.373977176s" podCreationTimestamp="2025-12-04 13:15:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:15:14.369747534 +0000 UTC m=+5538.644043338" watchObservedRunningTime="2025-12-04 13:15:14.373977176 +0000 UTC m=+5538.648272980" Dec 04 13:15:15 crc kubenswrapper[4979]: I1204 13:15:15.340560 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:17 crc kubenswrapper[4979]: I1204 13:15:17.515465 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:15:17 crc kubenswrapper[4979]: I1204 13:15:17.583962 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b79f8c4df-sc469"] Dec 04 13:15:17 crc kubenswrapper[4979]: I1204 13:15:17.584331 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" containerName="dnsmasq-dns" containerID="cri-o://c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885" gracePeriod=10 Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.107172 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.238665 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q59jj\" (UniqueName: \"kubernetes.io/projected/089bfb86-563e-4fea-88b7-845414efa9fe-kube-api-access-q59jj\") pod \"089bfb86-563e-4fea-88b7-845414efa9fe\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.238765 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-sb\") pod \"089bfb86-563e-4fea-88b7-845414efa9fe\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.238810 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-nb\") pod \"089bfb86-563e-4fea-88b7-845414efa9fe\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.238941 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-config\") pod \"089bfb86-563e-4fea-88b7-845414efa9fe\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.238975 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-dns-svc\") pod \"089bfb86-563e-4fea-88b7-845414efa9fe\" (UID: \"089bfb86-563e-4fea-88b7-845414efa9fe\") " Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.247642 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/089bfb86-563e-4fea-88b7-845414efa9fe-kube-api-access-q59jj" (OuterVolumeSpecName: "kube-api-access-q59jj") pod "089bfb86-563e-4fea-88b7-845414efa9fe" (UID: "089bfb86-563e-4fea-88b7-845414efa9fe"). InnerVolumeSpecName "kube-api-access-q59jj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.295569 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-config" (OuterVolumeSpecName: "config") pod "089bfb86-563e-4fea-88b7-845414efa9fe" (UID: "089bfb86-563e-4fea-88b7-845414efa9fe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.300963 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "089bfb86-563e-4fea-88b7-845414efa9fe" (UID: "089bfb86-563e-4fea-88b7-845414efa9fe"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.307062 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "089bfb86-563e-4fea-88b7-845414efa9fe" (UID: "089bfb86-563e-4fea-88b7-845414efa9fe"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.309973 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "089bfb86-563e-4fea-88b7-845414efa9fe" (UID: "089bfb86-563e-4fea-88b7-845414efa9fe"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.340975 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.341021 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.341033 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.341046 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q59jj\" (UniqueName: \"kubernetes.io/projected/089bfb86-563e-4fea-88b7-845414efa9fe-kube-api-access-q59jj\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.341058 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/089bfb86-563e-4fea-88b7-845414efa9fe-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.366934 4979 generic.go:334] "Generic (PLEG): container finished" podID="089bfb86-563e-4fea-88b7-845414efa9fe" containerID="c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885" exitCode=0 Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.366980 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" event={"ID":"089bfb86-563e-4fea-88b7-845414efa9fe","Type":"ContainerDied","Data":"c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885"} Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.366991 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.367005 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b79f8c4df-sc469" event={"ID":"089bfb86-563e-4fea-88b7-845414efa9fe","Type":"ContainerDied","Data":"38d25de7bad7b7510eb9353c9c03cfd9a83f885c11594b23d60436f83edbc433"} Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.367020 4979 scope.go:117] "RemoveContainer" containerID="c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.395977 4979 scope.go:117] "RemoveContainer" containerID="fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.414641 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b79f8c4df-sc469"] Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.429618 4979 scope.go:117] "RemoveContainer" containerID="c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885" Dec 04 13:15:18 crc kubenswrapper[4979]: E1204 13:15:18.430985 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885\": container with ID starting with c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885 not found: ID does not exist" containerID="c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.431037 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885"} err="failed to get container status \"c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885\": rpc error: code = NotFound desc = could not find container \"c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885\": container with ID starting with c911ab8bac014d8669ab9778fa6ba0ea1f963967357d9666428bfd603d9a9885 not found: ID does not exist" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.431085 4979 scope.go:117] "RemoveContainer" containerID="fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf" Dec 04 13:15:18 crc kubenswrapper[4979]: E1204 13:15:18.431738 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf\": container with ID starting with fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf not found: ID does not exist" containerID="fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.431760 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf"} err="failed to get container status \"fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf\": rpc error: code = NotFound desc = could not find container \"fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf\": container with ID starting with fdbed936a6b15aeee25b5803dfadb3e130ce396bc86cbb8c7de4e6bd38886dcf not found: ID does not exist" Dec 04 13:15:18 crc kubenswrapper[4979]: I1204 13:15:18.437467 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b79f8c4df-sc469"] Dec 04 13:15:20 crc kubenswrapper[4979]: I1204 13:15:20.210553 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" path="/var/lib/kubelet/pods/089bfb86-563e-4fea-88b7-845414efa9fe/volumes" Dec 04 13:15:43 crc kubenswrapper[4979]: I1204 13:15:43.846962 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:15:43 crc kubenswrapper[4979]: I1204 13:15:43.852086 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-69d694d89b-b86c2" Dec 04 13:16:06 crc kubenswrapper[4979]: I1204 13:16:06.762519 4979 scope.go:117] "RemoveContainer" containerID="bcd8476674daa4aad1246b52d39b39d1a9ab9763c05832179368c1b495b7aed0" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.267827 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-bf9ff"] Dec 04 13:16:08 crc kubenswrapper[4979]: E1204 13:16:08.268424 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" containerName="dnsmasq-dns" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.268435 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" containerName="dnsmasq-dns" Dec 04 13:16:08 crc kubenswrapper[4979]: E1204 13:16:08.268453 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" containerName="init" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.268459 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" containerName="init" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.268637 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="089bfb86-563e-4fea-88b7-845414efa9fe" containerName="dnsmasq-dns" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.269427 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.284493 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bf9ff"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.319849 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b71a0166-3ad9-47e6-9eec-7958a43af25c-operator-scripts\") pod \"nova-api-db-create-bf9ff\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.319905 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9g6b\" (UniqueName: \"kubernetes.io/projected/b71a0166-3ad9-47e6-9eec-7958a43af25c-kube-api-access-h9g6b\") pod \"nova-api-db-create-bf9ff\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.369317 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-cj9ww"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.370618 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.379489 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cj9ww"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.388572 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-6407-account-create-update-zl277"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.389898 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.391869 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.411564 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6407-account-create-update-zl277"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.421069 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9ndw\" (UniqueName: \"kubernetes.io/projected/0d3156be-8704-4b27-9cd9-1dd2b680d658-kube-api-access-l9ndw\") pod \"nova-api-6407-account-create-update-zl277\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.421126 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca30de84-d450-49e5-909e-fc83c3dc09ca-operator-scripts\") pod \"nova-cell0-db-create-cj9ww\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.421205 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d3156be-8704-4b27-9cd9-1dd2b680d658-operator-scripts\") pod \"nova-api-6407-account-create-update-zl277\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.421231 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnn8j\" (UniqueName: \"kubernetes.io/projected/ca30de84-d450-49e5-909e-fc83c3dc09ca-kube-api-access-hnn8j\") pod \"nova-cell0-db-create-cj9ww\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.421285 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b71a0166-3ad9-47e6-9eec-7958a43af25c-operator-scripts\") pod \"nova-api-db-create-bf9ff\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.421460 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9g6b\" (UniqueName: \"kubernetes.io/projected/b71a0166-3ad9-47e6-9eec-7958a43af25c-kube-api-access-h9g6b\") pod \"nova-api-db-create-bf9ff\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.422348 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b71a0166-3ad9-47e6-9eec-7958a43af25c-operator-scripts\") pod \"nova-api-db-create-bf9ff\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.443518 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9g6b\" (UniqueName: \"kubernetes.io/projected/b71a0166-3ad9-47e6-9eec-7958a43af25c-kube-api-access-h9g6b\") pod \"nova-api-db-create-bf9ff\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.474139 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-thbbq"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.475202 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.488357 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-thbbq"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.523636 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d3156be-8704-4b27-9cd9-1dd2b680d658-operator-scripts\") pod \"nova-api-6407-account-create-update-zl277\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.523692 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnn8j\" (UniqueName: \"kubernetes.io/projected/ca30de84-d450-49e5-909e-fc83c3dc09ca-kube-api-access-hnn8j\") pod \"nova-cell0-db-create-cj9ww\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.523770 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-operator-scripts\") pod \"nova-cell1-db-create-thbbq\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.523921 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhzgr\" (UniqueName: \"kubernetes.io/projected/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-kube-api-access-rhzgr\") pod \"nova-cell1-db-create-thbbq\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.523984 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9ndw\" (UniqueName: \"kubernetes.io/projected/0d3156be-8704-4b27-9cd9-1dd2b680d658-kube-api-access-l9ndw\") pod \"nova-api-6407-account-create-update-zl277\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.524020 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca30de84-d450-49e5-909e-fc83c3dc09ca-operator-scripts\") pod \"nova-cell0-db-create-cj9ww\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.524902 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d3156be-8704-4b27-9cd9-1dd2b680d658-operator-scripts\") pod \"nova-api-6407-account-create-update-zl277\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.525163 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca30de84-d450-49e5-909e-fc83c3dc09ca-operator-scripts\") pod \"nova-cell0-db-create-cj9ww\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.548445 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnn8j\" (UniqueName: \"kubernetes.io/projected/ca30de84-d450-49e5-909e-fc83c3dc09ca-kube-api-access-hnn8j\") pod \"nova-cell0-db-create-cj9ww\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.554909 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9ndw\" (UniqueName: \"kubernetes.io/projected/0d3156be-8704-4b27-9cd9-1dd2b680d658-kube-api-access-l9ndw\") pod \"nova-api-6407-account-create-update-zl277\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.573220 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9b7c-account-create-update-rb9m7"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.574260 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.576865 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.590537 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.597416 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b7c-account-create-update-rb9m7"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.626250 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-operator-scripts\") pod \"nova-cell1-db-create-thbbq\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.626762 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhzgr\" (UniqueName: \"kubernetes.io/projected/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-kube-api-access-rhzgr\") pod \"nova-cell1-db-create-thbbq\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.627132 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcqwx\" (UniqueName: \"kubernetes.io/projected/bc0a0593-111d-4346-a6e2-8831c97e3fd1-kube-api-access-mcqwx\") pod \"nova-cell0-9b7c-account-create-update-rb9m7\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.627202 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc0a0593-111d-4346-a6e2-8831c97e3fd1-operator-scripts\") pod \"nova-cell0-9b7c-account-create-update-rb9m7\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.627220 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-operator-scripts\") pod \"nova-cell1-db-create-thbbq\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.650413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhzgr\" (UniqueName: \"kubernetes.io/projected/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-kube-api-access-rhzgr\") pod \"nova-cell1-db-create-thbbq\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.689707 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.710397 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.728696 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcqwx\" (UniqueName: \"kubernetes.io/projected/bc0a0593-111d-4346-a6e2-8831c97e3fd1-kube-api-access-mcqwx\") pod \"nova-cell0-9b7c-account-create-update-rb9m7\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.728763 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc0a0593-111d-4346-a6e2-8831c97e3fd1-operator-scripts\") pod \"nova-cell0-9b7c-account-create-update-rb9m7\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.729513 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc0a0593-111d-4346-a6e2-8831c97e3fd1-operator-scripts\") pod \"nova-cell0-9b7c-account-create-update-rb9m7\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.766909 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcqwx\" (UniqueName: \"kubernetes.io/projected/bc0a0593-111d-4346-a6e2-8831c97e3fd1-kube-api-access-mcqwx\") pod \"nova-cell0-9b7c-account-create-update-rb9m7\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.779046 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.794853 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-292c-account-create-update-8l69r"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.796254 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.800784 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.815911 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-292c-account-create-update-8l69r"] Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.821151 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.834983 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhdh8\" (UniqueName: \"kubernetes.io/projected/e5bc4316-e6bc-4ab5-8d69-d45305523691-kube-api-access-qhdh8\") pod \"nova-cell1-292c-account-create-update-8l69r\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.839659 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5bc4316-e6bc-4ab5-8d69-d45305523691-operator-scripts\") pod \"nova-cell1-292c-account-create-update-8l69r\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.942403 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhdh8\" (UniqueName: \"kubernetes.io/projected/e5bc4316-e6bc-4ab5-8d69-d45305523691-kube-api-access-qhdh8\") pod \"nova-cell1-292c-account-create-update-8l69r\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.942532 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5bc4316-e6bc-4ab5-8d69-d45305523691-operator-scripts\") pod \"nova-cell1-292c-account-create-update-8l69r\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.943992 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5bc4316-e6bc-4ab5-8d69-d45305523691-operator-scripts\") pod \"nova-cell1-292c-account-create-update-8l69r\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:08 crc kubenswrapper[4979]: I1204 13:16:08.960804 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhdh8\" (UniqueName: \"kubernetes.io/projected/e5bc4316-e6bc-4ab5-8d69-d45305523691-kube-api-access-qhdh8\") pod \"nova-cell1-292c-account-create-update-8l69r\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:09 crc kubenswrapper[4979]: I1204 13:16:09.101469 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-bf9ff"] Dec 04 13:16:09 crc kubenswrapper[4979]: I1204 13:16:09.123212 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:09 crc kubenswrapper[4979]: I1204 13:16:09.275808 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-cj9ww"] Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.344267 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-6407-account-create-update-zl277"] Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.373777 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b7c-account-create-update-rb9m7"] Dec 04 13:16:10 crc kubenswrapper[4979]: W1204 13:16:09.375654 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc0a0593_111d_4346_a6e2_8831c97e3fd1.slice/crio-c46fdd56c29f12a93d10cd32fd2ce562f0d14427ce0ef257089e8b095e9a954a WatchSource:0}: Error finding container c46fdd56c29f12a93d10cd32fd2ce562f0d14427ce0ef257089e8b095e9a954a: Status 404 returned error can't find the container with id c46fdd56c29f12a93d10cd32fd2ce562f0d14427ce0ef257089e8b095e9a954a Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.403941 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-292c-account-create-update-8l69r"] Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.423778 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-thbbq"] Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.858847 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cj9ww" event={"ID":"ca30de84-d450-49e5-909e-fc83c3dc09ca","Type":"ContainerStarted","Data":"60dabd3bd55f05b03b087d547575cdd9816874a5885fc00f6b02fd600defe8f3"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.859252 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cj9ww" event={"ID":"ca30de84-d450-49e5-909e-fc83c3dc09ca","Type":"ContainerStarted","Data":"a876aef5366ea1fcc87aea678c9c1a58402f6bdf033559fe124dfda97869c20a"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.861248 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-292c-account-create-update-8l69r" event={"ID":"e5bc4316-e6bc-4ab5-8d69-d45305523691","Type":"ContainerStarted","Data":"a551fc0ac7677ef7085f345771fc062f2f63f7ef8c9068d2b14b1eb80e853ca8"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.870835 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6407-account-create-update-zl277" event={"ID":"0d3156be-8704-4b27-9cd9-1dd2b680d658","Type":"ContainerStarted","Data":"de67af96e04b099652a0b7a226661531dfd7351f084e569baaf093dd3248ba43"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.878761 4979 generic.go:334] "Generic (PLEG): container finished" podID="b71a0166-3ad9-47e6-9eec-7958a43af25c" containerID="2f32354589cfbd41f44ba0945614c186d20eb3b87d26b5b699fa1e67ab9eb088" exitCode=0 Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.878876 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bf9ff" event={"ID":"b71a0166-3ad9-47e6-9eec-7958a43af25c","Type":"ContainerDied","Data":"2f32354589cfbd41f44ba0945614c186d20eb3b87d26b5b699fa1e67ab9eb088"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.878909 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bf9ff" event={"ID":"b71a0166-3ad9-47e6-9eec-7958a43af25c","Type":"ContainerStarted","Data":"8894c3f78bdb98df242ace03f8ccea4dd3b8e32fda419637190b3cbcb90ac4e8"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.883752 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-thbbq" event={"ID":"f691ee1e-f296-4817-bc2d-35d0b7b10f7d","Type":"ContainerStarted","Data":"257c43c48b027672c5d9976155bca3bbf06bb93c8f0b28b49791db3a2a296400"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.887924 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-cj9ww" podStartSLOduration=1.887899411 podStartE2EDuration="1.887899411s" podCreationTimestamp="2025-12-04 13:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:09.877837804 +0000 UTC m=+5594.152133608" watchObservedRunningTime="2025-12-04 13:16:09.887899411 +0000 UTC m=+5594.162195215" Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.887989 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" event={"ID":"bc0a0593-111d-4346-a6e2-8831c97e3fd1","Type":"ContainerStarted","Data":"c46fdd56c29f12a93d10cd32fd2ce562f0d14427ce0ef257089e8b095e9a954a"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.920877 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-thbbq" podStartSLOduration=1.9208426360000002 podStartE2EDuration="1.920842636s" podCreationTimestamp="2025-12-04 13:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:09.917142339 +0000 UTC m=+5594.191438143" watchObservedRunningTime="2025-12-04 13:16:09.920842636 +0000 UTC m=+5594.195138440" Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:09.945534 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" podStartSLOduration=1.9455071830000001 podStartE2EDuration="1.945507183s" podCreationTimestamp="2025-12-04 13:16:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:09.932672381 +0000 UTC m=+5594.206968195" watchObservedRunningTime="2025-12-04 13:16:09.945507183 +0000 UTC m=+5594.219802987" Dec 04 13:16:10 crc kubenswrapper[4979]: E1204 13:16:10.243221 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf691ee1e_f296_4817_bc2d_35d0b7b10f7d.slice/crio-conmon-cab696639147898acb787288da495c7dd9c1889282a9371c77926c223c265eb2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf691ee1e_f296_4817_bc2d_35d0b7b10f7d.slice/crio-cab696639147898acb787288da495c7dd9c1889282a9371c77926c223c265eb2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5bc4316_e6bc_4ab5_8d69_d45305523691.slice/crio-conmon-8377b5b30088979b3a06fa5bb98286f97bf7c2e494558f84627ebb018bbf8c03.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5bc4316_e6bc_4ab5_8d69_d45305523691.slice/crio-8377b5b30088979b3a06fa5bb98286f97bf7c2e494558f84627ebb018bbf8c03.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d3156be_8704_4b27_9cd9_1dd2b680d658.slice/crio-3a3976ec5a3960ebe0310455ad3614b0c8ee93d0c921c736d3145b874d10140d.scope\": RecentStats: unable to find data in memory cache]" Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.899393 4979 generic.go:334] "Generic (PLEG): container finished" podID="f691ee1e-f296-4817-bc2d-35d0b7b10f7d" containerID="cab696639147898acb787288da495c7dd9c1889282a9371c77926c223c265eb2" exitCode=0 Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.899529 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-thbbq" event={"ID":"f691ee1e-f296-4817-bc2d-35d0b7b10f7d","Type":"ContainerDied","Data":"cab696639147898acb787288da495c7dd9c1889282a9371c77926c223c265eb2"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.901729 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc0a0593-111d-4346-a6e2-8831c97e3fd1" containerID="3b3b9a4fc820fbd1b5fd775bf777c40ce6981214ad55407d855291c8db502be8" exitCode=0 Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.901809 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" event={"ID":"bc0a0593-111d-4346-a6e2-8831c97e3fd1","Type":"ContainerDied","Data":"3b3b9a4fc820fbd1b5fd775bf777c40ce6981214ad55407d855291c8db502be8"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.904020 4979 generic.go:334] "Generic (PLEG): container finished" podID="ca30de84-d450-49e5-909e-fc83c3dc09ca" containerID="60dabd3bd55f05b03b087d547575cdd9816874a5885fc00f6b02fd600defe8f3" exitCode=0 Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.904063 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cj9ww" event={"ID":"ca30de84-d450-49e5-909e-fc83c3dc09ca","Type":"ContainerDied","Data":"60dabd3bd55f05b03b087d547575cdd9816874a5885fc00f6b02fd600defe8f3"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.908192 4979 generic.go:334] "Generic (PLEG): container finished" podID="e5bc4316-e6bc-4ab5-8d69-d45305523691" containerID="8377b5b30088979b3a06fa5bb98286f97bf7c2e494558f84627ebb018bbf8c03" exitCode=0 Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.908294 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-292c-account-create-update-8l69r" event={"ID":"e5bc4316-e6bc-4ab5-8d69-d45305523691","Type":"ContainerDied","Data":"8377b5b30088979b3a06fa5bb98286f97bf7c2e494558f84627ebb018bbf8c03"} Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.910240 4979 generic.go:334] "Generic (PLEG): container finished" podID="0d3156be-8704-4b27-9cd9-1dd2b680d658" containerID="3a3976ec5a3960ebe0310455ad3614b0c8ee93d0c921c736d3145b874d10140d" exitCode=0 Dec 04 13:16:10 crc kubenswrapper[4979]: I1204 13:16:10.910505 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6407-account-create-update-zl277" event={"ID":"0d3156be-8704-4b27-9cd9-1dd2b680d658","Type":"ContainerDied","Data":"3a3976ec5a3960ebe0310455ad3614b0c8ee93d0c921c736d3145b874d10140d"} Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.253893 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.293767 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9g6b\" (UniqueName: \"kubernetes.io/projected/b71a0166-3ad9-47e6-9eec-7958a43af25c-kube-api-access-h9g6b\") pod \"b71a0166-3ad9-47e6-9eec-7958a43af25c\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.293877 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b71a0166-3ad9-47e6-9eec-7958a43af25c-operator-scripts\") pod \"b71a0166-3ad9-47e6-9eec-7958a43af25c\" (UID: \"b71a0166-3ad9-47e6-9eec-7958a43af25c\") " Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.294857 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b71a0166-3ad9-47e6-9eec-7958a43af25c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b71a0166-3ad9-47e6-9eec-7958a43af25c" (UID: "b71a0166-3ad9-47e6-9eec-7958a43af25c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.302440 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b71a0166-3ad9-47e6-9eec-7958a43af25c-kube-api-access-h9g6b" (OuterVolumeSpecName: "kube-api-access-h9g6b") pod "b71a0166-3ad9-47e6-9eec-7958a43af25c" (UID: "b71a0166-3ad9-47e6-9eec-7958a43af25c"). InnerVolumeSpecName "kube-api-access-h9g6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.397256 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9g6b\" (UniqueName: \"kubernetes.io/projected/b71a0166-3ad9-47e6-9eec-7958a43af25c-kube-api-access-h9g6b\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.397341 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b71a0166-3ad9-47e6-9eec-7958a43af25c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.923713 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-bf9ff" Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.923710 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-bf9ff" event={"ID":"b71a0166-3ad9-47e6-9eec-7958a43af25c","Type":"ContainerDied","Data":"8894c3f78bdb98df242ace03f8ccea4dd3b8e32fda419637190b3cbcb90ac4e8"} Dec 04 13:16:11 crc kubenswrapper[4979]: I1204 13:16:11.923770 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8894c3f78bdb98df242ace03f8ccea4dd3b8e32fda419637190b3cbcb90ac4e8" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.309682 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.418983 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcqwx\" (UniqueName: \"kubernetes.io/projected/bc0a0593-111d-4346-a6e2-8831c97e3fd1-kube-api-access-mcqwx\") pod \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.419052 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc0a0593-111d-4346-a6e2-8831c97e3fd1-operator-scripts\") pod \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\" (UID: \"bc0a0593-111d-4346-a6e2-8831c97e3fd1\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.420151 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0a0593-111d-4346-a6e2-8831c97e3fd1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bc0a0593-111d-4346-a6e2-8831c97e3fd1" (UID: "bc0a0593-111d-4346-a6e2-8831c97e3fd1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.424315 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0a0593-111d-4346-a6e2-8831c97e3fd1-kube-api-access-mcqwx" (OuterVolumeSpecName: "kube-api-access-mcqwx") pod "bc0a0593-111d-4346-a6e2-8831c97e3fd1" (UID: "bc0a0593-111d-4346-a6e2-8831c97e3fd1"). InnerVolumeSpecName "kube-api-access-mcqwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.497364 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.509358 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.516225 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.521177 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhdh8\" (UniqueName: \"kubernetes.io/projected/e5bc4316-e6bc-4ab5-8d69-d45305523691-kube-api-access-qhdh8\") pod \"e5bc4316-e6bc-4ab5-8d69-d45305523691\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.521347 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5bc4316-e6bc-4ab5-8d69-d45305523691-operator-scripts\") pod \"e5bc4316-e6bc-4ab5-8d69-d45305523691\" (UID: \"e5bc4316-e6bc-4ab5-8d69-d45305523691\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.521698 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcqwx\" (UniqueName: \"kubernetes.io/projected/bc0a0593-111d-4346-a6e2-8831c97e3fd1-kube-api-access-mcqwx\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.521717 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bc0a0593-111d-4346-a6e2-8831c97e3fd1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.521857 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5bc4316-e6bc-4ab5-8d69-d45305523691-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5bc4316-e6bc-4ab5-8d69-d45305523691" (UID: "e5bc4316-e6bc-4ab5-8d69-d45305523691"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.523526 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.527679 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5bc4316-e6bc-4ab5-8d69-d45305523691-kube-api-access-qhdh8" (OuterVolumeSpecName: "kube-api-access-qhdh8") pod "e5bc4316-e6bc-4ab5-8d69-d45305523691" (UID: "e5bc4316-e6bc-4ab5-8d69-d45305523691"). InnerVolumeSpecName "kube-api-access-qhdh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.622882 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhzgr\" (UniqueName: \"kubernetes.io/projected/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-kube-api-access-rhzgr\") pod \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.622982 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnn8j\" (UniqueName: \"kubernetes.io/projected/ca30de84-d450-49e5-909e-fc83c3dc09ca-kube-api-access-hnn8j\") pod \"ca30de84-d450-49e5-909e-fc83c3dc09ca\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623023 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d3156be-8704-4b27-9cd9-1dd2b680d658-operator-scripts\") pod \"0d3156be-8704-4b27-9cd9-1dd2b680d658\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623100 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-operator-scripts\") pod \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\" (UID: \"f691ee1e-f296-4817-bc2d-35d0b7b10f7d\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623140 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9ndw\" (UniqueName: \"kubernetes.io/projected/0d3156be-8704-4b27-9cd9-1dd2b680d658-kube-api-access-l9ndw\") pod \"0d3156be-8704-4b27-9cd9-1dd2b680d658\" (UID: \"0d3156be-8704-4b27-9cd9-1dd2b680d658\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623168 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca30de84-d450-49e5-909e-fc83c3dc09ca-operator-scripts\") pod \"ca30de84-d450-49e5-909e-fc83c3dc09ca\" (UID: \"ca30de84-d450-49e5-909e-fc83c3dc09ca\") " Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623633 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhdh8\" (UniqueName: \"kubernetes.io/projected/e5bc4316-e6bc-4ab5-8d69-d45305523691-kube-api-access-qhdh8\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623656 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5bc4316-e6bc-4ab5-8d69-d45305523691-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623782 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f691ee1e-f296-4817-bc2d-35d0b7b10f7d" (UID: "f691ee1e-f296-4817-bc2d-35d0b7b10f7d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.623788 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d3156be-8704-4b27-9cd9-1dd2b680d658-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d3156be-8704-4b27-9cd9-1dd2b680d658" (UID: "0d3156be-8704-4b27-9cd9-1dd2b680d658"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.624026 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca30de84-d450-49e5-909e-fc83c3dc09ca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ca30de84-d450-49e5-909e-fc83c3dc09ca" (UID: "ca30de84-d450-49e5-909e-fc83c3dc09ca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.626072 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca30de84-d450-49e5-909e-fc83c3dc09ca-kube-api-access-hnn8j" (OuterVolumeSpecName: "kube-api-access-hnn8j") pod "ca30de84-d450-49e5-909e-fc83c3dc09ca" (UID: "ca30de84-d450-49e5-909e-fc83c3dc09ca"). InnerVolumeSpecName "kube-api-access-hnn8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.626108 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d3156be-8704-4b27-9cd9-1dd2b680d658-kube-api-access-l9ndw" (OuterVolumeSpecName: "kube-api-access-l9ndw") pod "0d3156be-8704-4b27-9cd9-1dd2b680d658" (UID: "0d3156be-8704-4b27-9cd9-1dd2b680d658"). InnerVolumeSpecName "kube-api-access-l9ndw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.626472 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-kube-api-access-rhzgr" (OuterVolumeSpecName: "kube-api-access-rhzgr") pod "f691ee1e-f296-4817-bc2d-35d0b7b10f7d" (UID: "f691ee1e-f296-4817-bc2d-35d0b7b10f7d"). InnerVolumeSpecName "kube-api-access-rhzgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.725454 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.725494 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9ndw\" (UniqueName: \"kubernetes.io/projected/0d3156be-8704-4b27-9cd9-1dd2b680d658-kube-api-access-l9ndw\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.725506 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ca30de84-d450-49e5-909e-fc83c3dc09ca-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.725514 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhzgr\" (UniqueName: \"kubernetes.io/projected/f691ee1e-f296-4817-bc2d-35d0b7b10f7d-kube-api-access-rhzgr\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.725523 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnn8j\" (UniqueName: \"kubernetes.io/projected/ca30de84-d450-49e5-909e-fc83c3dc09ca-kube-api-access-hnn8j\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.725532 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d3156be-8704-4b27-9cd9-1dd2b680d658-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.934004 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-cj9ww" event={"ID":"ca30de84-d450-49e5-909e-fc83c3dc09ca","Type":"ContainerDied","Data":"a876aef5366ea1fcc87aea678c9c1a58402f6bdf033559fe124dfda97869c20a"} Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.934027 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-cj9ww" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.934044 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a876aef5366ea1fcc87aea678c9c1a58402f6bdf033559fe124dfda97869c20a" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.935987 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-292c-account-create-update-8l69r" event={"ID":"e5bc4316-e6bc-4ab5-8d69-d45305523691","Type":"ContainerDied","Data":"a551fc0ac7677ef7085f345771fc062f2f63f7ef8c9068d2b14b1eb80e853ca8"} Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.936024 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a551fc0ac7677ef7085f345771fc062f2f63f7ef8c9068d2b14b1eb80e853ca8" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.935994 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-292c-account-create-update-8l69r" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.937930 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-6407-account-create-update-zl277" event={"ID":"0d3156be-8704-4b27-9cd9-1dd2b680d658","Type":"ContainerDied","Data":"de67af96e04b099652a0b7a226661531dfd7351f084e569baaf093dd3248ba43"} Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.937973 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de67af96e04b099652a0b7a226661531dfd7351f084e569baaf093dd3248ba43" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.938014 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-6407-account-create-update-zl277" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.940403 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-thbbq" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.940402 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-thbbq" event={"ID":"f691ee1e-f296-4817-bc2d-35d0b7b10f7d","Type":"ContainerDied","Data":"257c43c48b027672c5d9976155bca3bbf06bb93c8f0b28b49791db3a2a296400"} Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.940537 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="257c43c48b027672c5d9976155bca3bbf06bb93c8f0b28b49791db3a2a296400" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.942381 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" event={"ID":"bc0a0593-111d-4346-a6e2-8831c97e3fd1","Type":"ContainerDied","Data":"c46fdd56c29f12a93d10cd32fd2ce562f0d14427ce0ef257089e8b095e9a954a"} Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.942412 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c46fdd56c29f12a93d10cd32fd2ce562f0d14427ce0ef257089e8b095e9a954a" Dec 04 13:16:12 crc kubenswrapper[4979]: I1204 13:16:12.942595 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b7c-account-create-update-rb9m7" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.868507 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rjk2s"] Dec 04 13:16:13 crc kubenswrapper[4979]: E1204 13:16:13.869256 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f691ee1e-f296-4817-bc2d-35d0b7b10f7d" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869282 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f691ee1e-f296-4817-bc2d-35d0b7b10f7d" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: E1204 13:16:13.869320 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d3156be-8704-4b27-9cd9-1dd2b680d658" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869329 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d3156be-8704-4b27-9cd9-1dd2b680d658" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: E1204 13:16:13.869358 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0a0593-111d-4346-a6e2-8831c97e3fd1" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869368 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0a0593-111d-4346-a6e2-8831c97e3fd1" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: E1204 13:16:13.869388 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5bc4316-e6bc-4ab5-8d69-d45305523691" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869396 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5bc4316-e6bc-4ab5-8d69-d45305523691" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: E1204 13:16:13.869408 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b71a0166-3ad9-47e6-9eec-7958a43af25c" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869416 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b71a0166-3ad9-47e6-9eec-7958a43af25c" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: E1204 13:16:13.869451 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca30de84-d450-49e5-909e-fc83c3dc09ca" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869461 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca30de84-d450-49e5-909e-fc83c3dc09ca" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869696 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b71a0166-3ad9-47e6-9eec-7958a43af25c" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869720 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d3156be-8704-4b27-9cd9-1dd2b680d658" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869736 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f691ee1e-f296-4817-bc2d-35d0b7b10f7d" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869746 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5bc4316-e6bc-4ab5-8d69-d45305523691" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869768 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0a0593-111d-4346-a6e2-8831c97e3fd1" containerName="mariadb-account-create-update" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.869776 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca30de84-d450-49e5-909e-fc83c3dc09ca" containerName="mariadb-database-create" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.870622 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.873183 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.873557 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.875780 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-cjn76" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.885797 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rjk2s"] Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.945804 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-scripts\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.945926 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.945993 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmlvh\" (UniqueName: \"kubernetes.io/projected/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-kube-api-access-xmlvh\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:13 crc kubenswrapper[4979]: I1204 13:16:13.946026 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-config-data\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.047661 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.047753 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmlvh\" (UniqueName: \"kubernetes.io/projected/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-kube-api-access-xmlvh\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.047805 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-config-data\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.047887 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-scripts\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.053259 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.053347 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-scripts\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.053571 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-config-data\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.066193 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmlvh\" (UniqueName: \"kubernetes.io/projected/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-kube-api-access-xmlvh\") pod \"nova-cell0-conductor-db-sync-rjk2s\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.190667 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.657995 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rjk2s"] Dec 04 13:16:14 crc kubenswrapper[4979]: W1204 13:16:14.660116 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc98e14f3_b82e_41e7_8850_8ad4bfcb4b1f.slice/crio-911da0d3ba62a4b993995bc01045003d0da43d7e16b69af8f92e0fe6260b8870 WatchSource:0}: Error finding container 911da0d3ba62a4b993995bc01045003d0da43d7e16b69af8f92e0fe6260b8870: Status 404 returned error can't find the container with id 911da0d3ba62a4b993995bc01045003d0da43d7e16b69af8f92e0fe6260b8870 Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.961010 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" event={"ID":"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f","Type":"ContainerStarted","Data":"4caf6195a47109467c3b03032b264b2c37dd8d6870341f757c302d8879ecfd19"} Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.961346 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" event={"ID":"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f","Type":"ContainerStarted","Data":"911da0d3ba62a4b993995bc01045003d0da43d7e16b69af8f92e0fe6260b8870"} Dec 04 13:16:14 crc kubenswrapper[4979]: I1204 13:16:14.982241 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" podStartSLOduration=1.982220287 podStartE2EDuration="1.982220287s" podCreationTimestamp="2025-12-04 13:16:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:14.976420003 +0000 UTC m=+5599.250715807" watchObservedRunningTime="2025-12-04 13:16:14.982220287 +0000 UTC m=+5599.256516091" Dec 04 13:16:21 crc kubenswrapper[4979]: I1204 13:16:21.015264 4979 generic.go:334] "Generic (PLEG): container finished" podID="c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" containerID="4caf6195a47109467c3b03032b264b2c37dd8d6870341f757c302d8879ecfd19" exitCode=0 Dec 04 13:16:21 crc kubenswrapper[4979]: I1204 13:16:21.015360 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" event={"ID":"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f","Type":"ContainerDied","Data":"4caf6195a47109467c3b03032b264b2c37dd8d6870341f757c302d8879ecfd19"} Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.350655 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.497576 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmlvh\" (UniqueName: \"kubernetes.io/projected/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-kube-api-access-xmlvh\") pod \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.497714 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-combined-ca-bundle\") pod \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.497804 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-config-data\") pod \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.497879 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-scripts\") pod \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\" (UID: \"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f\") " Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.504733 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-kube-api-access-xmlvh" (OuterVolumeSpecName: "kube-api-access-xmlvh") pod "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" (UID: "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f"). InnerVolumeSpecName "kube-api-access-xmlvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.507982 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-scripts" (OuterVolumeSpecName: "scripts") pod "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" (UID: "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.524670 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-config-data" (OuterVolumeSpecName: "config-data") pod "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" (UID: "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.528834 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" (UID: "c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.600550 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmlvh\" (UniqueName: \"kubernetes.io/projected/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-kube-api-access-xmlvh\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.600623 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.600637 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:22 crc kubenswrapper[4979]: I1204 13:16:22.600649 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.035357 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" event={"ID":"c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f","Type":"ContainerDied","Data":"911da0d3ba62a4b993995bc01045003d0da43d7e16b69af8f92e0fe6260b8870"} Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.035675 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="911da0d3ba62a4b993995bc01045003d0da43d7e16b69af8f92e0fe6260b8870" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.035443 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-rjk2s" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.114515 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:16:23 crc kubenswrapper[4979]: E1204 13:16:23.114881 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" containerName="nova-cell0-conductor-db-sync" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.114896 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" containerName="nova-cell0-conductor-db-sync" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.115082 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" containerName="nova-cell0-conductor-db-sync" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.115786 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.117788 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-cjn76" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.118152 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.157000 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.212718 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.212871 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.212931 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9fxn\" (UniqueName: \"kubernetes.io/projected/83299115-b631-4e6e-901a-e82944e5c21c-kube-api-access-g9fxn\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.314951 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.315052 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9fxn\" (UniqueName: \"kubernetes.io/projected/83299115-b631-4e6e-901a-e82944e5c21c-kube-api-access-g9fxn\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.315384 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.320163 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.320540 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.336567 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9fxn\" (UniqueName: \"kubernetes.io/projected/83299115-b631-4e6e-901a-e82944e5c21c-kube-api-access-g9fxn\") pod \"nova-cell0-conductor-0\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.434685 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:23 crc kubenswrapper[4979]: W1204 13:16:23.864902 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83299115_b631_4e6e_901a_e82944e5c21c.slice/crio-e429d201d4cac11ffca9d562a17612ae4cf1c25103f97155aafd910072c1b584 WatchSource:0}: Error finding container e429d201d4cac11ffca9d562a17612ae4cf1c25103f97155aafd910072c1b584: Status 404 returned error can't find the container with id e429d201d4cac11ffca9d562a17612ae4cf1c25103f97155aafd910072c1b584 Dec 04 13:16:23 crc kubenswrapper[4979]: I1204 13:16:23.865239 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:16:24 crc kubenswrapper[4979]: I1204 13:16:24.044467 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"83299115-b631-4e6e-901a-e82944e5c21c","Type":"ContainerStarted","Data":"07bbe03959b56919a54c2cc12b07d0442d8ce804b250d0a6a5d7f7d887d3d15a"} Dec 04 13:16:24 crc kubenswrapper[4979]: I1204 13:16:24.044949 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"83299115-b631-4e6e-901a-e82944e5c21c","Type":"ContainerStarted","Data":"e429d201d4cac11ffca9d562a17612ae4cf1c25103f97155aafd910072c1b584"} Dec 04 13:16:24 crc kubenswrapper[4979]: I1204 13:16:24.044969 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:24 crc kubenswrapper[4979]: I1204 13:16:24.069089 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.0690641379999999 podStartE2EDuration="1.069064138s" podCreationTimestamp="2025-12-04 13:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:24.06046265 +0000 UTC m=+5608.334758464" watchObservedRunningTime="2025-12-04 13:16:24.069064138 +0000 UTC m=+5608.343359942" Dec 04 13:16:28 crc kubenswrapper[4979]: I1204 13:16:28.041616 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:16:28 crc kubenswrapper[4979]: I1204 13:16:28.043422 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:16:33 crc kubenswrapper[4979]: I1204 13:16:33.464929 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.035449 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-ljxl8"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.036671 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.046258 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ljxl8"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.047398 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.047595 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.166093 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.168062 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.172263 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.185470 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.186744 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.192648 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.214383 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.214473 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtvvf\" (UniqueName: \"kubernetes.io/projected/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-kube-api-access-mtvvf\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.214534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-config-data\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.214566 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-scripts\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.228181 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.228222 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.265357 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.266563 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.274339 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.277207 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315752 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42f4bed2-6142-4daf-b311-f1d7a718706d-logs\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315826 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315882 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcvxk\" (UniqueName: \"kubernetes.io/projected/249ad63a-955e-4fe2-b30a-4534dd49f8ef-kube-api-access-bcvxk\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315920 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtvvf\" (UniqueName: \"kubernetes.io/projected/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-kube-api-access-mtvvf\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315958 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315979 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.315997 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-config-data\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.316046 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-config-data\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.316071 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjmmr\" (UniqueName: \"kubernetes.io/projected/42f4bed2-6142-4daf-b311-f1d7a718706d-kube-api-access-wjmmr\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.316108 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-scripts\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.316149 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.328661 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-config-data\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.344438 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.350843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-scripts\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.352138 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.355164 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.356880 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtvvf\" (UniqueName: \"kubernetes.io/projected/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-kube-api-access-mtvvf\") pod \"nova-cell0-cell-mapping-ljxl8\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.362837 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.372524 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.417776 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-config-data\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.417844 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.417874 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42f4bed2-6142-4daf-b311-f1d7a718706d-logs\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.417930 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcvxk\" (UniqueName: \"kubernetes.io/projected/249ad63a-955e-4fe2-b30a-4534dd49f8ef-kube-api-access-bcvxk\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.417975 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.417997 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.418015 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-config-data\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.418051 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92dml\" (UniqueName: \"kubernetes.io/projected/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-kube-api-access-92dml\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.418076 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjmmr\" (UniqueName: \"kubernetes.io/projected/42f4bed2-6142-4daf-b311-f1d7a718706d-kube-api-access-wjmmr\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.418120 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.423811 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42f4bed2-6142-4daf-b311-f1d7a718706d-logs\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.426958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-config-data\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.429577 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.430631 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.434025 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.450333 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjmmr\" (UniqueName: \"kubernetes.io/projected/42f4bed2-6142-4daf-b311-f1d7a718706d-kube-api-access-wjmmr\") pod \"nova-api-0\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.455160 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcvxk\" (UniqueName: \"kubernetes.io/projected/249ad63a-955e-4fe2-b30a-4534dd49f8ef-kube-api-access-bcvxk\") pod \"nova-cell1-novncproxy-0\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.473575 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d598bd765-9z7pb"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.474975 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.493855 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.504416 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d598bd765-9z7pb"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92dml\" (UniqueName: \"kubernetes.io/projected/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-kube-api-access-92dml\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520686 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwzbt\" (UniqueName: \"kubernetes.io/projected/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-kube-api-access-nwzbt\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520748 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-config-data\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520769 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520806 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-logs\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520831 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.520858 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-config-data\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.524853 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.525171 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-config-data\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.525593 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.564344 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92dml\" (UniqueName: \"kubernetes.io/projected/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-kube-api-access-92dml\") pod \"nova-scheduler-0\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.593196 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.629811 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-nb\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630173 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-logs\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630208 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630241 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-config-data\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630385 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-sb\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwzbt\" (UniqueName: \"kubernetes.io/projected/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-kube-api-access-nwzbt\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630512 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-config\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630538 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-dns-svc\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.630581 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkrc8\" (UniqueName: \"kubernetes.io/projected/ce29f304-4be2-4e51-8d84-c0d20815a70c-kube-api-access-nkrc8\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.632534 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-logs\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.637042 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-config-data\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.637834 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.658936 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwzbt\" (UniqueName: \"kubernetes.io/projected/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-kube-api-access-nwzbt\") pod \"nova-metadata-0\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.658944 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.732426 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-config\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.732469 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-dns-svc\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.732500 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkrc8\" (UniqueName: \"kubernetes.io/projected/ce29f304-4be2-4e51-8d84-c0d20815a70c-kube-api-access-nkrc8\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.732539 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-nb\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.733564 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-sb\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.733736 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-nb\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.733763 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-dns-svc\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.733810 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-config\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.734282 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-sb\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.749058 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkrc8\" (UniqueName: \"kubernetes.io/projected/ce29f304-4be2-4e51-8d84-c0d20815a70c-kube-api-access-nkrc8\") pod \"dnsmasq-dns-6d598bd765-9z7pb\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.869041 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.889632 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.893010 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5c8l"] Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.894324 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.897646 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.903983 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 04 13:16:34 crc kubenswrapper[4979]: I1204 13:16:34.904904 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5c8l"] Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.049748 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvp5s\" (UniqueName: \"kubernetes.io/projected/e6741056-1468-4cc3-b55b-af86b1702693-kube-api-access-mvp5s\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.050163 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-config-data\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.050535 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.050738 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-scripts\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.060585 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.133524 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"249ad63a-955e-4fe2-b30a-4534dd49f8ef","Type":"ContainerStarted","Data":"903a00911c3dcf16f29a681e511d91d4cbd4e206f29321c2068e402004b339df"} Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.152441 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.152525 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-scripts\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.152596 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvp5s\" (UniqueName: \"kubernetes.io/projected/e6741056-1468-4cc3-b55b-af86b1702693-kube-api-access-mvp5s\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.152649 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-config-data\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.157997 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-scripts\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.158221 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-config-data\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.160838 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.176739 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvp5s\" (UniqueName: \"kubernetes.io/projected/e6741056-1468-4cc3-b55b-af86b1702693-kube-api-access-mvp5s\") pod \"nova-cell1-conductor-db-sync-s5c8l\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.180395 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.195658 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:35 crc kubenswrapper[4979]: W1204 13:16:35.196195 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42f4bed2_6142_4daf_b311_f1d7a718706d.slice/crio-6deb78db9953df5186876feb563bca64a6c00340645506c8d3f404b89e5c8825 WatchSource:0}: Error finding container 6deb78db9953df5186876feb563bca64a6c00340645506c8d3f404b89e5c8825: Status 404 returned error can't find the container with id 6deb78db9953df5186876feb563bca64a6c00340645506c8d3f404b89e5c8825 Dec 04 13:16:35 crc kubenswrapper[4979]: W1204 13:16:35.198407 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea4b6193_6341_4f23_9fc4_8ff1e04b6890.slice/crio-e26f67a302511ad0d78d9094992fdd9493332476ff8b65fc7e47287940d3bb4d WatchSource:0}: Error finding container e26f67a302511ad0d78d9094992fdd9493332476ff8b65fc7e47287940d3bb4d: Status 404 returned error can't find the container with id e26f67a302511ad0d78d9094992fdd9493332476ff8b65fc7e47287940d3bb4d Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.239214 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.268846 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-ljxl8"] Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.370409 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d598bd765-9z7pb"] Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.395489 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:35 crc kubenswrapper[4979]: I1204 13:16:35.743256 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5c8l"] Dec 04 13:16:35 crc kubenswrapper[4979]: W1204 13:16:35.754171 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6741056_1468_4cc3_b55b_af86b1702693.slice/crio-e1adad8a79fc572faeac660729cf42302ed5a4d9f2f8d15b88f6a2a9a9510cbd WatchSource:0}: Error finding container e1adad8a79fc572faeac660729cf42302ed5a4d9f2f8d15b88f6a2a9a9510cbd: Status 404 returned error can't find the container with id e1adad8a79fc572faeac660729cf42302ed5a4d9f2f8d15b88f6a2a9a9510cbd Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.158202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ea4b6193-6341-4f23-9fc4-8ff1e04b6890","Type":"ContainerStarted","Data":"eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.158643 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ea4b6193-6341-4f23-9fc4-8ff1e04b6890","Type":"ContainerStarted","Data":"e26f67a302511ad0d78d9094992fdd9493332476ff8b65fc7e47287940d3bb4d"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.162740 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" event={"ID":"e6741056-1468-4cc3-b55b-af86b1702693","Type":"ContainerStarted","Data":"83c1378a964ea7d546fb7e5044effc3da6d6be1e52c9ece03f3b751507a29b41"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.162790 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" event={"ID":"e6741056-1468-4cc3-b55b-af86b1702693","Type":"ContainerStarted","Data":"e1adad8a79fc572faeac660729cf42302ed5a4d9f2f8d15b88f6a2a9a9510cbd"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.165530 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42f4bed2-6142-4daf-b311-f1d7a718706d","Type":"ContainerStarted","Data":"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.165571 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42f4bed2-6142-4daf-b311-f1d7a718706d","Type":"ContainerStarted","Data":"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.165580 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42f4bed2-6142-4daf-b311-f1d7a718706d","Type":"ContainerStarted","Data":"6deb78db9953df5186876feb563bca64a6c00340645506c8d3f404b89e5c8825"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.168719 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"249ad63a-955e-4fe2-b30a-4534dd49f8ef","Type":"ContainerStarted","Data":"cb835efb2f8c182dc2aa148b9e0c0f19d2f55e14f952ef358c8d134fe28d9968"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.174727 4979 generic.go:334] "Generic (PLEG): container finished" podID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerID="ef777ec80b28541c1ecf9d5f2631f33bcf851102d1fdd8e3b8736f4db605fd18" exitCode=0 Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.174814 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" event={"ID":"ce29f304-4be2-4e51-8d84-c0d20815a70c","Type":"ContainerDied","Data":"ef777ec80b28541c1ecf9d5f2631f33bcf851102d1fdd8e3b8736f4db605fd18"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.174846 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" event={"ID":"ce29f304-4be2-4e51-8d84-c0d20815a70c","Type":"ContainerStarted","Data":"decc6b46a097403580ca08c689ef43d49902837f26dd1aa5428eb59ca57dd6fd"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.175285 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.175270576 podStartE2EDuration="2.175270576s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:36.174221958 +0000 UTC m=+5620.448517802" watchObservedRunningTime="2025-12-04 13:16:36.175270576 +0000 UTC m=+5620.449566380" Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.179154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f756044c-a2d8-4dbe-8849-a95f4c0eb52d","Type":"ContainerStarted","Data":"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.179205 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f756044c-a2d8-4dbe-8849-a95f4c0eb52d","Type":"ContainerStarted","Data":"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.179219 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f756044c-a2d8-4dbe-8849-a95f4c0eb52d","Type":"ContainerStarted","Data":"8f98d7abea15b03d19db87d33159593b4592f27d9dcda15862f0fe9c87979f84"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.183114 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ljxl8" event={"ID":"4d9fb02d-ba21-43b8-ba02-5d15d566d01e","Type":"ContainerStarted","Data":"e4bef2e6aea11e141ed1f01fed1205e41bd91b8727c00e92faf7dba24963fae9"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.183160 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ljxl8" event={"ID":"4d9fb02d-ba21-43b8-ba02-5d15d566d01e","Type":"ContainerStarted","Data":"74f2cfcf18bd3dc15910332cb93499a9790e57e5e89637ed8e7928a8ec6a0c17"} Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.225407 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.225387648 podStartE2EDuration="2.225387648s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:36.220022326 +0000 UTC m=+5620.494318150" watchObservedRunningTime="2025-12-04 13:16:36.225387648 +0000 UTC m=+5620.499683452" Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.230612 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" podStartSLOduration=2.230589247 podStartE2EDuration="2.230589247s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:36.19724806 +0000 UTC m=+5620.471543864" watchObservedRunningTime="2025-12-04 13:16:36.230589247 +0000 UTC m=+5620.504885051" Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.241790 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.241774404 podStartE2EDuration="2.241774404s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:36.233532035 +0000 UTC m=+5620.507827839" watchObservedRunningTime="2025-12-04 13:16:36.241774404 +0000 UTC m=+5620.516070208" Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.257210 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-ljxl8" podStartSLOduration=2.257190094 podStartE2EDuration="2.257190094s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:36.247335092 +0000 UTC m=+5620.521630886" watchObservedRunningTime="2025-12-04 13:16:36.257190094 +0000 UTC m=+5620.531485898" Dec 04 13:16:36 crc kubenswrapper[4979]: I1204 13:16:36.276549 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.276533438 podStartE2EDuration="2.276533438s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:36.265050483 +0000 UTC m=+5620.539346287" watchObservedRunningTime="2025-12-04 13:16:36.276533438 +0000 UTC m=+5620.550829232" Dec 04 13:16:37 crc kubenswrapper[4979]: I1204 13:16:37.195685 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" event={"ID":"ce29f304-4be2-4e51-8d84-c0d20815a70c","Type":"ContainerStarted","Data":"5a3ef93080fc9db8f226f566ab4397bc1ad4690bc98b9af793da0b10f60029a3"} Dec 04 13:16:37 crc kubenswrapper[4979]: I1204 13:16:37.196087 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:37 crc kubenswrapper[4979]: I1204 13:16:37.219718 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" podStartSLOduration=3.219700831 podStartE2EDuration="3.219700831s" podCreationTimestamp="2025-12-04 13:16:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:37.21701137 +0000 UTC m=+5621.491307174" watchObservedRunningTime="2025-12-04 13:16:37.219700831 +0000 UTC m=+5621.493996635" Dec 04 13:16:39 crc kubenswrapper[4979]: I1204 13:16:39.214251 4979 generic.go:334] "Generic (PLEG): container finished" podID="e6741056-1468-4cc3-b55b-af86b1702693" containerID="83c1378a964ea7d546fb7e5044effc3da6d6be1e52c9ece03f3b751507a29b41" exitCode=0 Dec 04 13:16:39 crc kubenswrapper[4979]: I1204 13:16:39.214357 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" event={"ID":"e6741056-1468-4cc3-b55b-af86b1702693","Type":"ContainerDied","Data":"83c1378a964ea7d546fb7e5044effc3da6d6be1e52c9ece03f3b751507a29b41"} Dec 04 13:16:39 crc kubenswrapper[4979]: I1204 13:16:39.526876 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:39 crc kubenswrapper[4979]: I1204 13:16:39.594189 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 13:16:39 crc kubenswrapper[4979]: I1204 13:16:39.870258 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:16:39 crc kubenswrapper[4979]: I1204 13:16:39.870371 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.632419 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.755750 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-config-data\") pod \"e6741056-1468-4cc3-b55b-af86b1702693\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.755860 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-combined-ca-bundle\") pod \"e6741056-1468-4cc3-b55b-af86b1702693\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.755965 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvp5s\" (UniqueName: \"kubernetes.io/projected/e6741056-1468-4cc3-b55b-af86b1702693-kube-api-access-mvp5s\") pod \"e6741056-1468-4cc3-b55b-af86b1702693\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.756068 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-scripts\") pod \"e6741056-1468-4cc3-b55b-af86b1702693\" (UID: \"e6741056-1468-4cc3-b55b-af86b1702693\") " Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.762708 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-scripts" (OuterVolumeSpecName: "scripts") pod "e6741056-1468-4cc3-b55b-af86b1702693" (UID: "e6741056-1468-4cc3-b55b-af86b1702693"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.763545 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6741056-1468-4cc3-b55b-af86b1702693-kube-api-access-mvp5s" (OuterVolumeSpecName: "kube-api-access-mvp5s") pod "e6741056-1468-4cc3-b55b-af86b1702693" (UID: "e6741056-1468-4cc3-b55b-af86b1702693"). InnerVolumeSpecName "kube-api-access-mvp5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.786158 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6741056-1468-4cc3-b55b-af86b1702693" (UID: "e6741056-1468-4cc3-b55b-af86b1702693"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.793466 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-config-data" (OuterVolumeSpecName: "config-data") pod "e6741056-1468-4cc3-b55b-af86b1702693" (UID: "e6741056-1468-4cc3-b55b-af86b1702693"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.857526 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.857689 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvp5s\" (UniqueName: \"kubernetes.io/projected/e6741056-1468-4cc3-b55b-af86b1702693-kube-api-access-mvp5s\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.857743 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:40 crc kubenswrapper[4979]: I1204 13:16:40.857793 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6741056-1468-4cc3-b55b-af86b1702693-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.241407 4979 generic.go:334] "Generic (PLEG): container finished" podID="4d9fb02d-ba21-43b8-ba02-5d15d566d01e" containerID="e4bef2e6aea11e141ed1f01fed1205e41bd91b8727c00e92faf7dba24963fae9" exitCode=0 Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.241926 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ljxl8" event={"ID":"4d9fb02d-ba21-43b8-ba02-5d15d566d01e","Type":"ContainerDied","Data":"e4bef2e6aea11e141ed1f01fed1205e41bd91b8727c00e92faf7dba24963fae9"} Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.249423 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" event={"ID":"e6741056-1468-4cc3-b55b-af86b1702693","Type":"ContainerDied","Data":"e1adad8a79fc572faeac660729cf42302ed5a4d9f2f8d15b88f6a2a9a9510cbd"} Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.249481 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1adad8a79fc572faeac660729cf42302ed5a4d9f2f8d15b88f6a2a9a9510cbd" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.249599 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-s5c8l" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.308720 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:16:41 crc kubenswrapper[4979]: E1204 13:16:41.309136 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6741056-1468-4cc3-b55b-af86b1702693" containerName="nova-cell1-conductor-db-sync" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.309158 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6741056-1468-4cc3-b55b-af86b1702693" containerName="nova-cell1-conductor-db-sync" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.309488 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6741056-1468-4cc3-b55b-af86b1702693" containerName="nova-cell1-conductor-db-sync" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.310855 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.314093 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.322111 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.439689 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.439828 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.439923 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk6rv\" (UniqueName: \"kubernetes.io/projected/2d0d52ad-7107-44f7-b317-d3b628ee078e-kube-api-access-hk6rv\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.541980 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.542062 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.542103 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk6rv\" (UniqueName: \"kubernetes.io/projected/2d0d52ad-7107-44f7-b317-d3b628ee078e-kube-api-access-hk6rv\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.546702 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.547190 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.562715 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk6rv\" (UniqueName: \"kubernetes.io/projected/2d0d52ad-7107-44f7-b317-d3b628ee078e-kube-api-access-hk6rv\") pod \"nova-cell1-conductor-0\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:41 crc kubenswrapper[4979]: I1204 13:16:41.633942 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.087959 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.259897 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0d52ad-7107-44f7-b317-d3b628ee078e","Type":"ContainerStarted","Data":"f3314bc7409f9cf1aaa693041c84d9f872ce68f7c3a376d966e3500eb87d6327"} Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.537220 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.666679 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-combined-ca-bundle\") pod \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.667030 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-config-data\") pod \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.667083 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtvvf\" (UniqueName: \"kubernetes.io/projected/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-kube-api-access-mtvvf\") pod \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.667193 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-scripts\") pod \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\" (UID: \"4d9fb02d-ba21-43b8-ba02-5d15d566d01e\") " Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.672952 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-scripts" (OuterVolumeSpecName: "scripts") pod "4d9fb02d-ba21-43b8-ba02-5d15d566d01e" (UID: "4d9fb02d-ba21-43b8-ba02-5d15d566d01e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.687720 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-kube-api-access-mtvvf" (OuterVolumeSpecName: "kube-api-access-mtvvf") pod "4d9fb02d-ba21-43b8-ba02-5d15d566d01e" (UID: "4d9fb02d-ba21-43b8-ba02-5d15d566d01e"). InnerVolumeSpecName "kube-api-access-mtvvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.696463 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-config-data" (OuterVolumeSpecName: "config-data") pod "4d9fb02d-ba21-43b8-ba02-5d15d566d01e" (UID: "4d9fb02d-ba21-43b8-ba02-5d15d566d01e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.709417 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d9fb02d-ba21-43b8-ba02-5d15d566d01e" (UID: "4d9fb02d-ba21-43b8-ba02-5d15d566d01e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.769851 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.769892 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.769908 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtvvf\" (UniqueName: \"kubernetes.io/projected/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-kube-api-access-mtvvf\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:42 crc kubenswrapper[4979]: I1204 13:16:42.769920 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4d9fb02d-ba21-43b8-ba02-5d15d566d01e-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.269696 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0d52ad-7107-44f7-b317-d3b628ee078e","Type":"ContainerStarted","Data":"e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03"} Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.270108 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.272268 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-ljxl8" event={"ID":"4d9fb02d-ba21-43b8-ba02-5d15d566d01e","Type":"ContainerDied","Data":"74f2cfcf18bd3dc15910332cb93499a9790e57e5e89637ed8e7928a8ec6a0c17"} Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.272353 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-ljxl8" Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.272389 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="74f2cfcf18bd3dc15910332cb93499a9790e57e5e89637ed8e7928a8ec6a0c17" Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.310995 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.310975329 podStartE2EDuration="2.310975329s" podCreationTimestamp="2025-12-04 13:16:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:43.293644679 +0000 UTC m=+5627.567940483" watchObservedRunningTime="2025-12-04 13:16:43.310975329 +0000 UTC m=+5627.585271133" Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.464620 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.464835 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ea4b6193-6341-4f23-9fc4-8ff1e04b6890" containerName="nova-scheduler-scheduler" containerID="cri-o://eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460" gracePeriod=30 Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.504354 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.504587 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-log" containerID="cri-o://f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5" gracePeriod=30 Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.505006 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-api" containerID="cri-o://e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee" gracePeriod=30 Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.513410 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.513646 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-log" containerID="cri-o://fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3" gracePeriod=30 Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.513789 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-metadata" containerID="cri-o://8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630" gracePeriod=30 Dec 04 13:16:43 crc kubenswrapper[4979]: I1204 13:16:43.957330 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.096542 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwzbt\" (UniqueName: \"kubernetes.io/projected/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-kube-api-access-nwzbt\") pod \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.096678 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-combined-ca-bundle\") pod \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.096716 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-config-data\") pod \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.096768 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-logs\") pod \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\" (UID: \"f756044c-a2d8-4dbe-8849-a95f4c0eb52d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.097804 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-logs" (OuterVolumeSpecName: "logs") pod "f756044c-a2d8-4dbe-8849-a95f4c0eb52d" (UID: "f756044c-a2d8-4dbe-8849-a95f4c0eb52d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.102328 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-kube-api-access-nwzbt" (OuterVolumeSpecName: "kube-api-access-nwzbt") pod "f756044c-a2d8-4dbe-8849-a95f4c0eb52d" (UID: "f756044c-a2d8-4dbe-8849-a95f4c0eb52d"). InnerVolumeSpecName "kube-api-access-nwzbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.112579 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.129931 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f756044c-a2d8-4dbe-8849-a95f4c0eb52d" (UID: "f756044c-a2d8-4dbe-8849-a95f4c0eb52d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.134061 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-config-data" (OuterVolumeSpecName: "config-data") pod "f756044c-a2d8-4dbe-8849-a95f4c0eb52d" (UID: "f756044c-a2d8-4dbe-8849-a95f4c0eb52d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.199027 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.199060 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwzbt\" (UniqueName: \"kubernetes.io/projected/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-kube-api-access-nwzbt\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.199070 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.199081 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f756044c-a2d8-4dbe-8849-a95f4c0eb52d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283641 4979 generic.go:334] "Generic (PLEG): container finished" podID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerID="8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630" exitCode=0 Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283687 4979 generic.go:334] "Generic (PLEG): container finished" podID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerID="fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3" exitCode=143 Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283693 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283748 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f756044c-a2d8-4dbe-8849-a95f4c0eb52d","Type":"ContainerDied","Data":"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630"} Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283797 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f756044c-a2d8-4dbe-8849-a95f4c0eb52d","Type":"ContainerDied","Data":"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3"} Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283811 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f756044c-a2d8-4dbe-8849-a95f4c0eb52d","Type":"ContainerDied","Data":"8f98d7abea15b03d19db87d33159593b4592f27d9dcda15862f0fe9c87979f84"} Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.283831 4979 scope.go:117] "RemoveContainer" containerID="8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.287427 4979 generic.go:334] "Generic (PLEG): container finished" podID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerID="e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee" exitCode=0 Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.287466 4979 generic.go:334] "Generic (PLEG): container finished" podID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerID="f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5" exitCode=143 Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.287572 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.288257 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42f4bed2-6142-4daf-b311-f1d7a718706d","Type":"ContainerDied","Data":"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee"} Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.288289 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42f4bed2-6142-4daf-b311-f1d7a718706d","Type":"ContainerDied","Data":"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5"} Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.288325 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"42f4bed2-6142-4daf-b311-f1d7a718706d","Type":"ContainerDied","Data":"6deb78db9953df5186876feb563bca64a6c00340645506c8d3f404b89e5c8825"} Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.304909 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjmmr\" (UniqueName: \"kubernetes.io/projected/42f4bed2-6142-4daf-b311-f1d7a718706d-kube-api-access-wjmmr\") pod \"42f4bed2-6142-4daf-b311-f1d7a718706d\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.305001 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-combined-ca-bundle\") pod \"42f4bed2-6142-4daf-b311-f1d7a718706d\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.305032 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42f4bed2-6142-4daf-b311-f1d7a718706d-logs\") pod \"42f4bed2-6142-4daf-b311-f1d7a718706d\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.305070 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-config-data\") pod \"42f4bed2-6142-4daf-b311-f1d7a718706d\" (UID: \"42f4bed2-6142-4daf-b311-f1d7a718706d\") " Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.305787 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42f4bed2-6142-4daf-b311-f1d7a718706d-logs" (OuterVolumeSpecName: "logs") pod "42f4bed2-6142-4daf-b311-f1d7a718706d" (UID: "42f4bed2-6142-4daf-b311-f1d7a718706d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.305907 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42f4bed2-6142-4daf-b311-f1d7a718706d-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.309527 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42f4bed2-6142-4daf-b311-f1d7a718706d-kube-api-access-wjmmr" (OuterVolumeSpecName: "kube-api-access-wjmmr") pod "42f4bed2-6142-4daf-b311-f1d7a718706d" (UID: "42f4bed2-6142-4daf-b311-f1d7a718706d"). InnerVolumeSpecName "kube-api-access-wjmmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.313419 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.317375 4979 scope.go:117] "RemoveContainer" containerID="fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.353489 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-config-data" (OuterVolumeSpecName: "config-data") pod "42f4bed2-6142-4daf-b311-f1d7a718706d" (UID: "42f4bed2-6142-4daf-b311-f1d7a718706d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.357924 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.362257 4979 scope.go:117] "RemoveContainer" containerID="8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.362783 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630\": container with ID starting with 8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630 not found: ID does not exist" containerID="8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.362813 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630"} err="failed to get container status \"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630\": rpc error: code = NotFound desc = could not find container \"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630\": container with ID starting with 8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630 not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.362835 4979 scope.go:117] "RemoveContainer" containerID="fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.366854 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3\": container with ID starting with fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3 not found: ID does not exist" containerID="fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.367052 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3"} err="failed to get container status \"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3\": rpc error: code = NotFound desc = could not find container \"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3\": container with ID starting with fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3 not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.367156 4979 scope.go:117] "RemoveContainer" containerID="8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.372223 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630"} err="failed to get container status \"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630\": rpc error: code = NotFound desc = could not find container \"8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630\": container with ID starting with 8353c6006784eeec45bc7a9c18eebe7597459c2aba49c62a553f0a6d9f0ed630 not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.372281 4979 scope.go:117] "RemoveContainer" containerID="fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373115 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373238 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42f4bed2-6142-4daf-b311-f1d7a718706d" (UID: "42f4bed2-6142-4daf-b311-f1d7a718706d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.373624 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-log" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373645 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-log" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.373678 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-api" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373686 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-api" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.373712 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-metadata" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373720 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-metadata" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.373737 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d9fb02d-ba21-43b8-ba02-5d15d566d01e" containerName="nova-manage" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373745 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d9fb02d-ba21-43b8-ba02-5d15d566d01e" containerName="nova-manage" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.373755 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-log" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373762 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-log" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373973 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d9fb02d-ba21-43b8-ba02-5d15d566d01e" containerName="nova-manage" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.373996 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-log" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.374015 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-metadata" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.374025 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" containerName="nova-metadata-log" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.374041 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" containerName="nova-api-api" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.374757 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3"} err="failed to get container status \"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3\": rpc error: code = NotFound desc = could not find container \"fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3\": container with ID starting with fbc3a6ec358f474731427719e9df64d241671e29836f46a88dfcea8ed903f4d3 not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.374854 4979 scope.go:117] "RemoveContainer" containerID="e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.376488 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.379143 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.386637 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.404034 4979 scope.go:117] "RemoveContainer" containerID="f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407042 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1476130-876c-4b1b-a638-d1fdf3ef22d7-logs\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407111 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407222 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njqsb\" (UniqueName: \"kubernetes.io/projected/f1476130-876c-4b1b-a638-d1fdf3ef22d7-kube-api-access-njqsb\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407514 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407784 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjmmr\" (UniqueName: \"kubernetes.io/projected/42f4bed2-6142-4daf-b311-f1d7a718706d-kube-api-access-wjmmr\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407803 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.407811 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42f4bed2-6142-4daf-b311-f1d7a718706d-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.422852 4979 scope.go:117] "RemoveContainer" containerID="e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.424123 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee\": container with ID starting with e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee not found: ID does not exist" containerID="e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.424200 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee"} err="failed to get container status \"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee\": rpc error: code = NotFound desc = could not find container \"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee\": container with ID starting with e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.424230 4979 scope.go:117] "RemoveContainer" containerID="f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5" Dec 04 13:16:44 crc kubenswrapper[4979]: E1204 13:16:44.424783 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5\": container with ID starting with f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5 not found: ID does not exist" containerID="f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.424822 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5"} err="failed to get container status \"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5\": rpc error: code = NotFound desc = could not find container \"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5\": container with ID starting with f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5 not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.424853 4979 scope.go:117] "RemoveContainer" containerID="e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.425423 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee"} err="failed to get container status \"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee\": rpc error: code = NotFound desc = could not find container \"e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee\": container with ID starting with e492751575eee16e5c79670289493b480290c9f36a361cd256a427c3d88832ee not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.425501 4979 scope.go:117] "RemoveContainer" containerID="f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.425983 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5"} err="failed to get container status \"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5\": rpc error: code = NotFound desc = could not find container \"f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5\": container with ID starting with f9c90e35168869182f94d8a0856b59ea8aba82b56f7e4dc68b5452c5d7bab7c5 not found: ID does not exist" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.509145 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.509286 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1476130-876c-4b1b-a638-d1fdf3ef22d7-logs\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.509352 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.509396 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njqsb\" (UniqueName: \"kubernetes.io/projected/f1476130-876c-4b1b-a638-d1fdf3ef22d7-kube-api-access-njqsb\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.509980 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1476130-876c-4b1b-a638-d1fdf3ef22d7-logs\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.513027 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.514358 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.526339 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.529735 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njqsb\" (UniqueName: \"kubernetes.io/projected/f1476130-876c-4b1b-a638-d1fdf3ef22d7-kube-api-access-njqsb\") pod \"nova-metadata-0\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.537530 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.627052 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.638268 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.652741 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.654430 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.657706 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.668925 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.697213 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.712636 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dbb1680-0d7e-4922-8e09-bc030dfb498f-logs\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.712715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9t4w\" (UniqueName: \"kubernetes.io/projected/3dbb1680-0d7e-4922-8e09-bc030dfb498f-kube-api-access-z9t4w\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.712768 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.712813 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-config-data\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.814232 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dbb1680-0d7e-4922-8e09-bc030dfb498f-logs\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.814693 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9t4w\" (UniqueName: \"kubernetes.io/projected/3dbb1680-0d7e-4922-8e09-bc030dfb498f-kube-api-access-z9t4w\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.814713 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dbb1680-0d7e-4922-8e09-bc030dfb498f-logs\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.814789 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.814867 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-config-data\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.820349 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.825324 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-config-data\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.833432 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9t4w\" (UniqueName: \"kubernetes.io/projected/3dbb1680-0d7e-4922-8e09-bc030dfb498f-kube-api-access-z9t4w\") pod \"nova-api-0\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " pod="openstack/nova-api-0" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.892704 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.961535 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9db9c4bf-l6bqz"] Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.961834 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" podUID="1174334f-0570-4363-b95b-3870ca0ace66" containerName="dnsmasq-dns" containerID="cri-o://dd5acd3226af73f6fddcaea2f639962db5478838e225ca36a89084474d352db4" gracePeriod=10 Dec 04 13:16:44 crc kubenswrapper[4979]: I1204 13:16:44.998395 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.142771 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.303985 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1476130-876c-4b1b-a638-d1fdf3ef22d7","Type":"ContainerStarted","Data":"d26204e48f1c5211ff2047384ac5fd25caf64676bad8f38a440f8c7e23646737"} Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.310228 4979 generic.go:334] "Generic (PLEG): container finished" podID="1174334f-0570-4363-b95b-3870ca0ace66" containerID="dd5acd3226af73f6fddcaea2f639962db5478838e225ca36a89084474d352db4" exitCode=0 Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.310372 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" event={"ID":"1174334f-0570-4363-b95b-3870ca0ace66","Type":"ContainerDied","Data":"dd5acd3226af73f6fddcaea2f639962db5478838e225ca36a89084474d352db4"} Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.322703 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.463586 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:16:45 crc kubenswrapper[4979]: W1204 13:16:45.465703 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3dbb1680_0d7e_4922_8e09_bc030dfb498f.slice/crio-f6ff213168ed29a38a16dd723762ff29ec2dc89948b1439c0825828102e7faa4 WatchSource:0}: Error finding container f6ff213168ed29a38a16dd723762ff29ec2dc89948b1439c0825828102e7faa4: Status 404 returned error can't find the container with id f6ff213168ed29a38a16dd723762ff29ec2dc89948b1439c0825828102e7faa4 Dec 04 13:16:45 crc kubenswrapper[4979]: I1204 13:16:45.918053 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.042758 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mggj\" (UniqueName: \"kubernetes.io/projected/1174334f-0570-4363-b95b-3870ca0ace66-kube-api-access-4mggj\") pod \"1174334f-0570-4363-b95b-3870ca0ace66\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.042826 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-config\") pod \"1174334f-0570-4363-b95b-3870ca0ace66\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.042914 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-dns-svc\") pod \"1174334f-0570-4363-b95b-3870ca0ace66\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.042965 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-sb\") pod \"1174334f-0570-4363-b95b-3870ca0ace66\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.043041 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-nb\") pod \"1174334f-0570-4363-b95b-3870ca0ace66\" (UID: \"1174334f-0570-4363-b95b-3870ca0ace66\") " Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.048611 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1174334f-0570-4363-b95b-3870ca0ace66-kube-api-access-4mggj" (OuterVolumeSpecName: "kube-api-access-4mggj") pod "1174334f-0570-4363-b95b-3870ca0ace66" (UID: "1174334f-0570-4363-b95b-3870ca0ace66"). InnerVolumeSpecName "kube-api-access-4mggj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.097591 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1174334f-0570-4363-b95b-3870ca0ace66" (UID: "1174334f-0570-4363-b95b-3870ca0ace66"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.097658 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1174334f-0570-4363-b95b-3870ca0ace66" (UID: "1174334f-0570-4363-b95b-3870ca0ace66"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.099184 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-config" (OuterVolumeSpecName: "config") pod "1174334f-0570-4363-b95b-3870ca0ace66" (UID: "1174334f-0570-4363-b95b-3870ca0ace66"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.103361 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1174334f-0570-4363-b95b-3870ca0ace66" (UID: "1174334f-0570-4363-b95b-3870ca0ace66"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.144615 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mggj\" (UniqueName: \"kubernetes.io/projected/1174334f-0570-4363-b95b-3870ca0ace66-kube-api-access-4mggj\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.144924 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.144933 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.144945 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.144954 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1174334f-0570-4363-b95b-3870ca0ace66-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.210587 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42f4bed2-6142-4daf-b311-f1d7a718706d" path="/var/lib/kubelet/pods/42f4bed2-6142-4daf-b311-f1d7a718706d/volumes" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.211244 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f756044c-a2d8-4dbe-8849-a95f4c0eb52d" path="/var/lib/kubelet/pods/f756044c-a2d8-4dbe-8849-a95f4c0eb52d/volumes" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.331454 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1476130-876c-4b1b-a638-d1fdf3ef22d7","Type":"ContainerStarted","Data":"adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c"} Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.331511 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1476130-876c-4b1b-a638-d1fdf3ef22d7","Type":"ContainerStarted","Data":"ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07"} Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.334321 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" event={"ID":"1174334f-0570-4363-b95b-3870ca0ace66","Type":"ContainerDied","Data":"a0ba8a4ab6bf6dcc03d00d61a2040cf2298e367758166088c47363085515bb89"} Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.334367 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9db9c4bf-l6bqz" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.334454 4979 scope.go:117] "RemoveContainer" containerID="dd5acd3226af73f6fddcaea2f639962db5478838e225ca36a89084474d352db4" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.336625 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dbb1680-0d7e-4922-8e09-bc030dfb498f","Type":"ContainerStarted","Data":"128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056"} Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.336669 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dbb1680-0d7e-4922-8e09-bc030dfb498f","Type":"ContainerStarted","Data":"980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92"} Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.336680 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dbb1680-0d7e-4922-8e09-bc030dfb498f","Type":"ContainerStarted","Data":"f6ff213168ed29a38a16dd723762ff29ec2dc89948b1439c0825828102e7faa4"} Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.360690 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.360672281 podStartE2EDuration="2.360672281s" podCreationTimestamp="2025-12-04 13:16:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:46.350532212 +0000 UTC m=+5630.624828036" watchObservedRunningTime="2025-12-04 13:16:46.360672281 +0000 UTC m=+5630.634968085" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.360796 4979 scope.go:117] "RemoveContainer" containerID="3a6ed20a59ba4903324ea0804508a1173ae9d76988d18e4229ee3848b5d98f07" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.383603 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.383581581 podStartE2EDuration="2.383581581s" podCreationTimestamp="2025-12-04 13:16:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:46.37526788 +0000 UTC m=+5630.649563684" watchObservedRunningTime="2025-12-04 13:16:46.383581581 +0000 UTC m=+5630.657877385" Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.401762 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9db9c4bf-l6bqz"] Dec 04 13:16:46 crc kubenswrapper[4979]: I1204 13:16:46.423048 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9db9c4bf-l6bqz"] Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.111218 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.263675 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92dml\" (UniqueName: \"kubernetes.io/projected/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-kube-api-access-92dml\") pod \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.263779 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-config-data\") pod \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.263855 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-combined-ca-bundle\") pod \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\" (UID: \"ea4b6193-6341-4f23-9fc4-8ff1e04b6890\") " Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.269828 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-kube-api-access-92dml" (OuterVolumeSpecName: "kube-api-access-92dml") pod "ea4b6193-6341-4f23-9fc4-8ff1e04b6890" (UID: "ea4b6193-6341-4f23-9fc4-8ff1e04b6890"). InnerVolumeSpecName "kube-api-access-92dml". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.292025 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ea4b6193-6341-4f23-9fc4-8ff1e04b6890" (UID: "ea4b6193-6341-4f23-9fc4-8ff1e04b6890"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.297103 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-config-data" (OuterVolumeSpecName: "config-data") pod "ea4b6193-6341-4f23-9fc4-8ff1e04b6890" (UID: "ea4b6193-6341-4f23-9fc4-8ff1e04b6890"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.350732 4979 generic.go:334] "Generic (PLEG): container finished" podID="ea4b6193-6341-4f23-9fc4-8ff1e04b6890" containerID="eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460" exitCode=0 Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.350880 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ea4b6193-6341-4f23-9fc4-8ff1e04b6890","Type":"ContainerDied","Data":"eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460"} Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.350937 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ea4b6193-6341-4f23-9fc4-8ff1e04b6890","Type":"ContainerDied","Data":"e26f67a302511ad0d78d9094992fdd9493332476ff8b65fc7e47287940d3bb4d"} Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.350959 4979 scope.go:117] "RemoveContainer" containerID="eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.351115 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.366879 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92dml\" (UniqueName: \"kubernetes.io/projected/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-kube-api-access-92dml\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.366908 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.366917 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea4b6193-6341-4f23-9fc4-8ff1e04b6890-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.373455 4979 scope.go:117] "RemoveContainer" containerID="eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460" Dec 04 13:16:47 crc kubenswrapper[4979]: E1204 13:16:47.373869 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460\": container with ID starting with eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460 not found: ID does not exist" containerID="eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.373903 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460"} err="failed to get container status \"eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460\": rpc error: code = NotFound desc = could not find container \"eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460\": container with ID starting with eca547913d3c9d5eb5c987af1d25970db702956f25b863fce82956f7ee64f460 not found: ID does not exist" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.390125 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.402360 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.412866 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:47 crc kubenswrapper[4979]: E1204 13:16:47.413332 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1174334f-0570-4363-b95b-3870ca0ace66" containerName="init" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.413350 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1174334f-0570-4363-b95b-3870ca0ace66" containerName="init" Dec 04 13:16:47 crc kubenswrapper[4979]: E1204 13:16:47.413364 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1174334f-0570-4363-b95b-3870ca0ace66" containerName="dnsmasq-dns" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.413370 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1174334f-0570-4363-b95b-3870ca0ace66" containerName="dnsmasq-dns" Dec 04 13:16:47 crc kubenswrapper[4979]: E1204 13:16:47.413391 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea4b6193-6341-4f23-9fc4-8ff1e04b6890" containerName="nova-scheduler-scheduler" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.413397 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea4b6193-6341-4f23-9fc4-8ff1e04b6890" containerName="nova-scheduler-scheduler" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.413574 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea4b6193-6341-4f23-9fc4-8ff1e04b6890" containerName="nova-scheduler-scheduler" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.413605 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1174334f-0570-4363-b95b-3870ca0ace66" containerName="dnsmasq-dns" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.414322 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.416569 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.428818 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.570269 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.570376 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-config-data\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.570935 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cx5s\" (UniqueName: \"kubernetes.io/projected/eeb01655-e25d-4250-934e-3bf24f8a808c-kube-api-access-9cx5s\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.672647 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cx5s\" (UniqueName: \"kubernetes.io/projected/eeb01655-e25d-4250-934e-3bf24f8a808c-kube-api-access-9cx5s\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.672997 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.673136 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-config-data\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.676782 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.677981 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-config-data\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.690921 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cx5s\" (UniqueName: \"kubernetes.io/projected/eeb01655-e25d-4250-934e-3bf24f8a808c-kube-api-access-9cx5s\") pod \"nova-scheduler-0\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " pod="openstack/nova-scheduler-0" Dec 04 13:16:47 crc kubenswrapper[4979]: I1204 13:16:47.744097 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:16:48 crc kubenswrapper[4979]: I1204 13:16:48.169150 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:16:48 crc kubenswrapper[4979]: W1204 13:16:48.171902 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeeb01655_e25d_4250_934e_3bf24f8a808c.slice/crio-8d0eade3f3a9fdb4f871b28b635d95bab0eb75d9bd7c07bfb0d74c2bffd184c9 WatchSource:0}: Error finding container 8d0eade3f3a9fdb4f871b28b635d95bab0eb75d9bd7c07bfb0d74c2bffd184c9: Status 404 returned error can't find the container with id 8d0eade3f3a9fdb4f871b28b635d95bab0eb75d9bd7c07bfb0d74c2bffd184c9 Dec 04 13:16:48 crc kubenswrapper[4979]: I1204 13:16:48.211475 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1174334f-0570-4363-b95b-3870ca0ace66" path="/var/lib/kubelet/pods/1174334f-0570-4363-b95b-3870ca0ace66/volumes" Dec 04 13:16:48 crc kubenswrapper[4979]: I1204 13:16:48.212225 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea4b6193-6341-4f23-9fc4-8ff1e04b6890" path="/var/lib/kubelet/pods/ea4b6193-6341-4f23-9fc4-8ff1e04b6890/volumes" Dec 04 13:16:48 crc kubenswrapper[4979]: I1204 13:16:48.365781 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb01655-e25d-4250-934e-3bf24f8a808c","Type":"ContainerStarted","Data":"8d0eade3f3a9fdb4f871b28b635d95bab0eb75d9bd7c07bfb0d74c2bffd184c9"} Dec 04 13:16:49 crc kubenswrapper[4979]: I1204 13:16:49.381070 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb01655-e25d-4250-934e-3bf24f8a808c","Type":"ContainerStarted","Data":"a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05"} Dec 04 13:16:49 crc kubenswrapper[4979]: I1204 13:16:49.401590 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.40156993 podStartE2EDuration="2.40156993s" podCreationTimestamp="2025-12-04 13:16:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:49.396987998 +0000 UTC m=+5633.671283832" watchObservedRunningTime="2025-12-04 13:16:49.40156993 +0000 UTC m=+5633.675865734" Dec 04 13:16:49 crc kubenswrapper[4979]: I1204 13:16:49.697715 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:16:49 crc kubenswrapper[4979]: I1204 13:16:49.697775 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:16:51 crc kubenswrapper[4979]: I1204 13:16:51.663518 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.071344 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-gpkvf"] Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.072738 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.084619 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gpkvf"] Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.098242 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.098669 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.254351 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mz58\" (UniqueName: \"kubernetes.io/projected/f9f3c3d0-6cfd-4b30-b836-62671e12f561-kube-api-access-8mz58\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.254434 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-scripts\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.254614 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-config-data\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.254781 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.356687 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.356776 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mz58\" (UniqueName: \"kubernetes.io/projected/f9f3c3d0-6cfd-4b30-b836-62671e12f561-kube-api-access-8mz58\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.356835 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-scripts\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.356910 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-config-data\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.362823 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-config-data\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.362837 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.362992 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-scripts\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.375645 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mz58\" (UniqueName: \"kubernetes.io/projected/f9f3c3d0-6cfd-4b30-b836-62671e12f561-kube-api-access-8mz58\") pod \"nova-cell1-cell-mapping-gpkvf\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.398109 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.745182 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 13:16:52 crc kubenswrapper[4979]: I1204 13:16:52.831578 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-gpkvf"] Dec 04 13:16:53 crc kubenswrapper[4979]: I1204 13:16:53.413565 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gpkvf" event={"ID":"f9f3c3d0-6cfd-4b30-b836-62671e12f561","Type":"ContainerStarted","Data":"b44a1ad0450092e02a5bbaea44d48fa05dc01d8d7df47246441925799ab8bfb4"} Dec 04 13:16:53 crc kubenswrapper[4979]: I1204 13:16:53.413625 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gpkvf" event={"ID":"f9f3c3d0-6cfd-4b30-b836-62671e12f561","Type":"ContainerStarted","Data":"4dec059a3941a577ce135060db43894100893075f5e4b29aee1ff986cfd51e9b"} Dec 04 13:16:53 crc kubenswrapper[4979]: I1204 13:16:53.436782 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-gpkvf" podStartSLOduration=1.43676308 podStartE2EDuration="1.43676308s" podCreationTimestamp="2025-12-04 13:16:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:16:53.428283525 +0000 UTC m=+5637.702579329" watchObservedRunningTime="2025-12-04 13:16:53.43676308 +0000 UTC m=+5637.711058884" Dec 04 13:16:54 crc kubenswrapper[4979]: I1204 13:16:54.698326 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 13:16:54 crc kubenswrapper[4979]: I1204 13:16:54.701233 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 13:16:54 crc kubenswrapper[4979]: I1204 13:16:54.999340 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 13:16:54 crc kubenswrapper[4979]: I1204 13:16:54.999869 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 13:16:55 crc kubenswrapper[4979]: I1204 13:16:55.781527 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:16:55 crc kubenswrapper[4979]: I1204 13:16:55.781986 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.69:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:16:56 crc kubenswrapper[4979]: I1204 13:16:56.082549 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:16:56 crc kubenswrapper[4979]: I1204 13:16:56.082538 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.70:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:16:57 crc kubenswrapper[4979]: I1204 13:16:57.744600 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 13:16:57 crc kubenswrapper[4979]: I1204 13:16:57.772333 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 13:16:58 crc kubenswrapper[4979]: I1204 13:16:58.040731 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:16:58 crc kubenswrapper[4979]: I1204 13:16:58.040782 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:16:58 crc kubenswrapper[4979]: I1204 13:16:58.482385 4979 generic.go:334] "Generic (PLEG): container finished" podID="f9f3c3d0-6cfd-4b30-b836-62671e12f561" containerID="b44a1ad0450092e02a5bbaea44d48fa05dc01d8d7df47246441925799ab8bfb4" exitCode=0 Dec 04 13:16:58 crc kubenswrapper[4979]: I1204 13:16:58.482496 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gpkvf" event={"ID":"f9f3c3d0-6cfd-4b30-b836-62671e12f561","Type":"ContainerDied","Data":"b44a1ad0450092e02a5bbaea44d48fa05dc01d8d7df47246441925799ab8bfb4"} Dec 04 13:16:58 crc kubenswrapper[4979]: I1204 13:16:58.513367 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.813771 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.895995 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-combined-ca-bundle\") pod \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.896097 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mz58\" (UniqueName: \"kubernetes.io/projected/f9f3c3d0-6cfd-4b30-b836-62671e12f561-kube-api-access-8mz58\") pod \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.896251 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-config-data\") pod \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.896276 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-scripts\") pod \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\" (UID: \"f9f3c3d0-6cfd-4b30-b836-62671e12f561\") " Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.901880 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-scripts" (OuterVolumeSpecName: "scripts") pod "f9f3c3d0-6cfd-4b30-b836-62671e12f561" (UID: "f9f3c3d0-6cfd-4b30-b836-62671e12f561"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.901958 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9f3c3d0-6cfd-4b30-b836-62671e12f561-kube-api-access-8mz58" (OuterVolumeSpecName: "kube-api-access-8mz58") pod "f9f3c3d0-6cfd-4b30-b836-62671e12f561" (UID: "f9f3c3d0-6cfd-4b30-b836-62671e12f561"). InnerVolumeSpecName "kube-api-access-8mz58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.921834 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-config-data" (OuterVolumeSpecName: "config-data") pod "f9f3c3d0-6cfd-4b30-b836-62671e12f561" (UID: "f9f3c3d0-6cfd-4b30-b836-62671e12f561"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.922765 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9f3c3d0-6cfd-4b30-b836-62671e12f561" (UID: "f9f3c3d0-6cfd-4b30-b836-62671e12f561"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.998849 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.998889 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.998898 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9f3c3d0-6cfd-4b30-b836-62671e12f561-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:16:59 crc kubenswrapper[4979]: I1204 13:16:59.998909 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mz58\" (UniqueName: \"kubernetes.io/projected/f9f3c3d0-6cfd-4b30-b836-62671e12f561-kube-api-access-8mz58\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.501923 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-gpkvf" event={"ID":"f9f3c3d0-6cfd-4b30-b836-62671e12f561","Type":"ContainerDied","Data":"4dec059a3941a577ce135060db43894100893075f5e4b29aee1ff986cfd51e9b"} Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.502245 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4dec059a3941a577ce135060db43894100893075f5e4b29aee1ff986cfd51e9b" Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.501942 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-gpkvf" Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.675694 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.675939 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-log" containerID="cri-o://980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92" gracePeriod=30 Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.676064 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-api" containerID="cri-o://128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056" gracePeriod=30 Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.709925 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.710128 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="eeb01655-e25d-4250-934e-3bf24f8a808c" containerName="nova-scheduler-scheduler" containerID="cri-o://a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" gracePeriod=30 Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.732650 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.732917 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-log" containerID="cri-o://ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07" gracePeriod=30 Dec 04 13:17:00 crc kubenswrapper[4979]: I1204 13:17:00.733082 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-metadata" containerID="cri-o://adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c" gracePeriod=30 Dec 04 13:17:01 crc kubenswrapper[4979]: I1204 13:17:01.511193 4979 generic.go:334] "Generic (PLEG): container finished" podID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerID="ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07" exitCode=143 Dec 04 13:17:01 crc kubenswrapper[4979]: I1204 13:17:01.511283 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1476130-876c-4b1b-a638-d1fdf3ef22d7","Type":"ContainerDied","Data":"ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07"} Dec 04 13:17:01 crc kubenswrapper[4979]: I1204 13:17:01.517785 4979 generic.go:334] "Generic (PLEG): container finished" podID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerID="980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92" exitCode=143 Dec 04 13:17:01 crc kubenswrapper[4979]: I1204 13:17:01.517823 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dbb1680-0d7e-4922-8e09-bc030dfb498f","Type":"ContainerDied","Data":"980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92"} Dec 04 13:17:02 crc kubenswrapper[4979]: E1204 13:17:02.746840 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 13:17:02 crc kubenswrapper[4979]: E1204 13:17:02.748230 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 13:17:02 crc kubenswrapper[4979]: E1204 13:17:02.749497 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 13:17:02 crc kubenswrapper[4979]: E1204 13:17:02.749606 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="eeb01655-e25d-4250-934e-3bf24f8a808c" containerName="nova-scheduler-scheduler" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.255471 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.320845 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.376726 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-config-data\") pod \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.376879 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9t4w\" (UniqueName: \"kubernetes.io/projected/3dbb1680-0d7e-4922-8e09-bc030dfb498f-kube-api-access-z9t4w\") pod \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.376947 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-combined-ca-bundle\") pod \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.376995 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dbb1680-0d7e-4922-8e09-bc030dfb498f-logs\") pod \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\" (UID: \"3dbb1680-0d7e-4922-8e09-bc030dfb498f\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.377605 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dbb1680-0d7e-4922-8e09-bc030dfb498f-logs" (OuterVolumeSpecName: "logs") pod "3dbb1680-0d7e-4922-8e09-bc030dfb498f" (UID: "3dbb1680-0d7e-4922-8e09-bc030dfb498f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.382287 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dbb1680-0d7e-4922-8e09-bc030dfb498f-kube-api-access-z9t4w" (OuterVolumeSpecName: "kube-api-access-z9t4w") pod "3dbb1680-0d7e-4922-8e09-bc030dfb498f" (UID: "3dbb1680-0d7e-4922-8e09-bc030dfb498f"). InnerVolumeSpecName "kube-api-access-z9t4w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.405292 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-config-data" (OuterVolumeSpecName: "config-data") pod "3dbb1680-0d7e-4922-8e09-bc030dfb498f" (UID: "3dbb1680-0d7e-4922-8e09-bc030dfb498f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.407973 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dbb1680-0d7e-4922-8e09-bc030dfb498f" (UID: "3dbb1680-0d7e-4922-8e09-bc030dfb498f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.478896 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1476130-876c-4b1b-a638-d1fdf3ef22d7-logs\") pod \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479063 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njqsb\" (UniqueName: \"kubernetes.io/projected/f1476130-876c-4b1b-a638-d1fdf3ef22d7-kube-api-access-njqsb\") pod \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479192 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-combined-ca-bundle\") pod \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479230 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data\") pod \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479718 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479740 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9t4w\" (UniqueName: \"kubernetes.io/projected/3dbb1680-0d7e-4922-8e09-bc030dfb498f-kube-api-access-z9t4w\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479754 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dbb1680-0d7e-4922-8e09-bc030dfb498f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.479763 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dbb1680-0d7e-4922-8e09-bc030dfb498f-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.480742 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1476130-876c-4b1b-a638-d1fdf3ef22d7-logs" (OuterVolumeSpecName: "logs") pod "f1476130-876c-4b1b-a638-d1fdf3ef22d7" (UID: "f1476130-876c-4b1b-a638-d1fdf3ef22d7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.483526 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1476130-876c-4b1b-a638-d1fdf3ef22d7-kube-api-access-njqsb" (OuterVolumeSpecName: "kube-api-access-njqsb") pod "f1476130-876c-4b1b-a638-d1fdf3ef22d7" (UID: "f1476130-876c-4b1b-a638-d1fdf3ef22d7"). InnerVolumeSpecName "kube-api-access-njqsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.500345 4979 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data podName:f1476130-876c-4b1b-a638-d1fdf3ef22d7 nodeName:}" failed. No retries permitted until 2025-12-04 13:17:05.00031303 +0000 UTC m=+5649.274608844 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data") pod "f1476130-876c-4b1b-a638-d1fdf3ef22d7" (UID: "f1476130-876c-4b1b-a638-d1fdf3ef22d7") : error deleting /var/lib/kubelet/pods/f1476130-876c-4b1b-a638-d1fdf3ef22d7/volume-subpaths: remove /var/lib/kubelet/pods/f1476130-876c-4b1b-a638-d1fdf3ef22d7/volume-subpaths: no such file or directory Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.502701 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1476130-876c-4b1b-a638-d1fdf3ef22d7" (UID: "f1476130-876c-4b1b-a638-d1fdf3ef22d7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.544130 4979 generic.go:334] "Generic (PLEG): container finished" podID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerID="adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c" exitCode=0 Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.544192 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.544200 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1476130-876c-4b1b-a638-d1fdf3ef22d7","Type":"ContainerDied","Data":"adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c"} Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.544651 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f1476130-876c-4b1b-a638-d1fdf3ef22d7","Type":"ContainerDied","Data":"d26204e48f1c5211ff2047384ac5fd25caf64676bad8f38a440f8c7e23646737"} Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.544674 4979 scope.go:117] "RemoveContainer" containerID="adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.546147 4979 generic.go:334] "Generic (PLEG): container finished" podID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerID="128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056" exitCode=0 Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.546204 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.546314 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dbb1680-0d7e-4922-8e09-bc030dfb498f","Type":"ContainerDied","Data":"128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056"} Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.546491 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dbb1680-0d7e-4922-8e09-bc030dfb498f","Type":"ContainerDied","Data":"f6ff213168ed29a38a16dd723762ff29ec2dc89948b1439c0825828102e7faa4"} Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.569433 4979 scope.go:117] "RemoveContainer" containerID="ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.581101 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.581132 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1476130-876c-4b1b-a638-d1fdf3ef22d7-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.581142 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njqsb\" (UniqueName: \"kubernetes.io/projected/f1476130-876c-4b1b-a638-d1fdf3ef22d7-kube-api-access-njqsb\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.592679 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.600172 4979 scope.go:117] "RemoveContainer" containerID="adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.601666 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c\": container with ID starting with adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c not found: ID does not exist" containerID="adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.601711 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c"} err="failed to get container status \"adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c\": rpc error: code = NotFound desc = could not find container \"adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c\": container with ID starting with adc8ae14c74789f3a04a2b17128d89fa235ba0ca4284b1b25dbbe6b10e06954c not found: ID does not exist" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.601746 4979 scope.go:117] "RemoveContainer" containerID="ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.603560 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07\": container with ID starting with ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07 not found: ID does not exist" containerID="ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.603592 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07"} err="failed to get container status \"ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07\": rpc error: code = NotFound desc = could not find container \"ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07\": container with ID starting with ae325586b01f91aec919f362b1d069dd816ee89a4311cd89df73be0fc03c2c07 not found: ID does not exist" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.603611 4979 scope.go:117] "RemoveContainer" containerID="128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.607702 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.627718 4979 scope.go:117] "RemoveContainer" containerID="980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.642691 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.643188 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-api" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643208 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-api" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.643224 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-log" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643231 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-log" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.643250 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-log" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643259 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-log" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.643283 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-metadata" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643290 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-metadata" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.643342 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9f3c3d0-6cfd-4b30-b836-62671e12f561" containerName="nova-manage" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643352 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9f3c3d0-6cfd-4b30-b836-62671e12f561" containerName="nova-manage" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643552 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-log" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643564 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9f3c3d0-6cfd-4b30-b836-62671e12f561" containerName="nova-manage" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643577 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" containerName="nova-metadata-metadata" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643592 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-log" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.643612 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" containerName="nova-api-api" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.645012 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.650966 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.655619 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.661132 4979 scope.go:117] "RemoveContainer" containerID="128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.661991 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056\": container with ID starting with 128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056 not found: ID does not exist" containerID="128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.662016 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056"} err="failed to get container status \"128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056\": rpc error: code = NotFound desc = could not find container \"128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056\": container with ID starting with 128ecb08f2f364ad1cac1ced198ce011f212c410afb7ce201d2765ffc5aba056 not found: ID does not exist" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.662061 4979 scope.go:117] "RemoveContainer" containerID="980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92" Dec 04 13:17:04 crc kubenswrapper[4979]: E1204 13:17:04.662412 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92\": container with ID starting with 980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92 not found: ID does not exist" containerID="980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.662445 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92"} err="failed to get container status \"980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92\": rpc error: code = NotFound desc = could not find container \"980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92\": container with ID starting with 980f04b32964efed94a81180d91d7352d3262d6f6f1866640a2cb09347388a92 not found: ID does not exist" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.784204 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbtpl\" (UniqueName: \"kubernetes.io/projected/e28a7327-643f-45f1-848c-d971f083136a-kube-api-access-gbtpl\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.784413 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-config-data\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.784458 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28a7327-643f-45f1-848c-d971f083136a-logs\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.784474 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.886163 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-config-data\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.886229 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28a7327-643f-45f1-848c-d971f083136a-logs\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.886243 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.886285 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbtpl\" (UniqueName: \"kubernetes.io/projected/e28a7327-643f-45f1-848c-d971f083136a-kube-api-access-gbtpl\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.887175 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28a7327-643f-45f1-848c-d971f083136a-logs\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.890589 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-config-data\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.892625 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.903734 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbtpl\" (UniqueName: \"kubernetes.io/projected/e28a7327-643f-45f1-848c-d971f083136a-kube-api-access-gbtpl\") pod \"nova-api-0\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " pod="openstack/nova-api-0" Dec 04 13:17:04 crc kubenswrapper[4979]: I1204 13:17:04.968596 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.073677 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.088345 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data\") pod \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\" (UID: \"f1476130-876c-4b1b-a638-d1fdf3ef22d7\") " Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.093436 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data" (OuterVolumeSpecName: "config-data") pod "f1476130-876c-4b1b-a638-d1fdf3ef22d7" (UID: "f1476130-876c-4b1b-a638-d1fdf3ef22d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.178838 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.189840 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-config-data\") pod \"eeb01655-e25d-4250-934e-3bf24f8a808c\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.190026 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cx5s\" (UniqueName: \"kubernetes.io/projected/eeb01655-e25d-4250-934e-3bf24f8a808c-kube-api-access-9cx5s\") pod \"eeb01655-e25d-4250-934e-3bf24f8a808c\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.190057 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-combined-ca-bundle\") pod \"eeb01655-e25d-4250-934e-3bf24f8a808c\" (UID: \"eeb01655-e25d-4250-934e-3bf24f8a808c\") " Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.190760 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.191548 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1476130-876c-4b1b-a638-d1fdf3ef22d7-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.194986 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eeb01655-e25d-4250-934e-3bf24f8a808c-kube-api-access-9cx5s" (OuterVolumeSpecName: "kube-api-access-9cx5s") pod "eeb01655-e25d-4250-934e-3bf24f8a808c" (UID: "eeb01655-e25d-4250-934e-3bf24f8a808c"). InnerVolumeSpecName "kube-api-access-9cx5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.209353 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: E1204 13:17:05.209916 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eeb01655-e25d-4250-934e-3bf24f8a808c" containerName="nova-scheduler-scheduler" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.209982 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="eeb01655-e25d-4250-934e-3bf24f8a808c" containerName="nova-scheduler-scheduler" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.210269 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="eeb01655-e25d-4250-934e-3bf24f8a808c" containerName="nova-scheduler-scheduler" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.211777 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.214454 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.228981 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.238030 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eeb01655-e25d-4250-934e-3bf24f8a808c" (UID: "eeb01655-e25d-4250-934e-3bf24f8a808c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.270422 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-config-data" (OuterVolumeSpecName: "config-data") pod "eeb01655-e25d-4250-934e-3bf24f8a808c" (UID: "eeb01655-e25d-4250-934e-3bf24f8a808c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.293239 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.293630 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-config-data\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.293716 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8l6g\" (UniqueName: \"kubernetes.io/projected/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-kube-api-access-x8l6g\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.293821 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-logs\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.293946 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cx5s\" (UniqueName: \"kubernetes.io/projected/eeb01655-e25d-4250-934e-3bf24f8a808c-kube-api-access-9cx5s\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.294009 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.294082 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eeb01655-e25d-4250-934e-3bf24f8a808c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.396060 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.396139 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-config-data\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.396205 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8l6g\" (UniqueName: \"kubernetes.io/projected/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-kube-api-access-x8l6g\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.396231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-logs\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.396882 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-logs\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.401268 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.401550 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-config-data\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: W1204 13:17:05.408590 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode28a7327_643f_45f1_848c_d971f083136a.slice/crio-e13eb49d8bf633856d3ba3faae306a08cf119b4046907dbf13f8e96ef79c814b WatchSource:0}: Error finding container e13eb49d8bf633856d3ba3faae306a08cf119b4046907dbf13f8e96ef79c814b: Status 404 returned error can't find the container with id e13eb49d8bf633856d3ba3faae306a08cf119b4046907dbf13f8e96ef79c814b Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.409196 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.414266 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8l6g\" (UniqueName: \"kubernetes.io/projected/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-kube-api-access-x8l6g\") pod \"nova-metadata-0\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.556860 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e28a7327-643f-45f1-848c-d971f083136a","Type":"ContainerStarted","Data":"e13eb49d8bf633856d3ba3faae306a08cf119b4046907dbf13f8e96ef79c814b"} Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.557074 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.560901 4979 generic.go:334] "Generic (PLEG): container finished" podID="eeb01655-e25d-4250-934e-3bf24f8a808c" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" exitCode=0 Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.560942 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb01655-e25d-4250-934e-3bf24f8a808c","Type":"ContainerDied","Data":"a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05"} Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.560974 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"eeb01655-e25d-4250-934e-3bf24f8a808c","Type":"ContainerDied","Data":"8d0eade3f3a9fdb4f871b28b635d95bab0eb75d9bd7c07bfb0d74c2bffd184c9"} Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.560983 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.560990 4979 scope.go:117] "RemoveContainer" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.589807 4979 scope.go:117] "RemoveContainer" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" Dec 04 13:17:05 crc kubenswrapper[4979]: E1204 13:17:05.590286 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05\": container with ID starting with a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05 not found: ID does not exist" containerID="a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.590464 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05"} err="failed to get container status \"a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05\": rpc error: code = NotFound desc = could not find container \"a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05\": container with ID starting with a2daad55101c305cafa10807deb24beca06901c114eb55c61d037d0e7bb9ca05 not found: ID does not exist" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.598609 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.622499 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.631889 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.633506 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.639872 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.643679 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.804362 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nblch\" (UniqueName: \"kubernetes.io/projected/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-kube-api-access-nblch\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.804517 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.804664 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-config-data\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.906478 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nblch\" (UniqueName: \"kubernetes.io/projected/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-kube-api-access-nblch\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.906620 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.907396 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-config-data\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.911271 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.912793 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-config-data\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.923069 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nblch\" (UniqueName: \"kubernetes.io/projected/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-kube-api-access-nblch\") pod \"nova-scheduler-0\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " pod="openstack/nova-scheduler-0" Dec 04 13:17:05 crc kubenswrapper[4979]: I1204 13:17:05.960242 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.035380 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:17:06 crc kubenswrapper[4979]: W1204 13:17:06.039147 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbda865f5_85ba_4d30_84e7_057bc6e2e1c9.slice/crio-f575bba795b0e8ea12b20ba1c206db9ebbe7a27ec7509a7011f1708888b8bf79 WatchSource:0}: Error finding container f575bba795b0e8ea12b20ba1c206db9ebbe7a27ec7509a7011f1708888b8bf79: Status 404 returned error can't find the container with id f575bba795b0e8ea12b20ba1c206db9ebbe7a27ec7509a7011f1708888b8bf79 Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.216016 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dbb1680-0d7e-4922-8e09-bc030dfb498f" path="/var/lib/kubelet/pods/3dbb1680-0d7e-4922-8e09-bc030dfb498f/volumes" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.217378 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eeb01655-e25d-4250-934e-3bf24f8a808c" path="/var/lib/kubelet/pods/eeb01655-e25d-4250-934e-3bf24f8a808c/volumes" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.218039 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1476130-876c-4b1b-a638-d1fdf3ef22d7" path="/var/lib/kubelet/pods/f1476130-876c-4b1b-a638-d1fdf3ef22d7/volumes" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.419327 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:17:06 crc kubenswrapper[4979]: W1204 13:17:06.429450 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod665c8ff9_2c3b_42e1_8cc4_dd0f18b5e3c5.slice/crio-d3002c585acdc293a4cbe1f144e8c0e91190cfe1c3f3cb0e5c24723ae8f7a0a1 WatchSource:0}: Error finding container d3002c585acdc293a4cbe1f144e8c0e91190cfe1c3f3cb0e5c24723ae8f7a0a1: Status 404 returned error can't find the container with id d3002c585acdc293a4cbe1f144e8c0e91190cfe1c3f3cb0e5c24723ae8f7a0a1 Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.574488 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e28a7327-643f-45f1-848c-d971f083136a","Type":"ContainerStarted","Data":"6aca61a38fb477a9267cb4b0138b2346e65ac917289490a6259f0aff4c6f5466"} Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.574958 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e28a7327-643f-45f1-848c-d971f083136a","Type":"ContainerStarted","Data":"8e6430c1953b1184982391fbe8caa201a15a1e05432d4e5e5b07895d2aa2f71d"} Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.577625 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5","Type":"ContainerStarted","Data":"d3002c585acdc293a4cbe1f144e8c0e91190cfe1c3f3cb0e5c24723ae8f7a0a1"} Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.583449 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bda865f5-85ba-4d30-84e7-057bc6e2e1c9","Type":"ContainerStarted","Data":"4c9a940c9aefdc19229ce43bd16b871d1dbc397f9bbd271044a98121efd9b57c"} Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.583483 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bda865f5-85ba-4d30-84e7-057bc6e2e1c9","Type":"ContainerStarted","Data":"1da2e815fda862ad0ddafcd951202e197e5c13bd9f123d32d3ec9bc4c29e8edd"} Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.583496 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bda865f5-85ba-4d30-84e7-057bc6e2e1c9","Type":"ContainerStarted","Data":"f575bba795b0e8ea12b20ba1c206db9ebbe7a27ec7509a7011f1708888b8bf79"} Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.624470 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.624449307 podStartE2EDuration="1.624449307s" podCreationTimestamp="2025-12-04 13:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:06.61892312 +0000 UTC m=+5650.893218924" watchObservedRunningTime="2025-12-04 13:17:06.624449307 +0000 UTC m=+5650.898745111" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.625073 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.625064983 podStartE2EDuration="2.625064983s" podCreationTimestamp="2025-12-04 13:17:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:06.60045567 +0000 UTC m=+5650.874751474" watchObservedRunningTime="2025-12-04 13:17:06.625064983 +0000 UTC m=+5650.899360787" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.831133 4979 scope.go:117] "RemoveContainer" containerID="8af1e4510309f447a9af3a1f372710c1c495d92cb253ea23e18a59d5bb24a7f2" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.870933 4979 scope.go:117] "RemoveContainer" containerID="6cfa180fcc4b2f4c636eba9b60cb0273ebee86ce6190eea95f74161b0a9fde06" Dec 04 13:17:06 crc kubenswrapper[4979]: I1204 13:17:06.897465 4979 scope.go:117] "RemoveContainer" containerID="a1b27d160ea952de7f5474594e3538fb6f02ad6224f6fd9996061e2e25e596ac" Dec 04 13:17:07 crc kubenswrapper[4979]: I1204 13:17:07.597224 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5","Type":"ContainerStarted","Data":"4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d"} Dec 04 13:17:07 crc kubenswrapper[4979]: I1204 13:17:07.623721 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.62370232 podStartE2EDuration="2.62370232s" podCreationTimestamp="2025-12-04 13:17:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:07.622440018 +0000 UTC m=+5651.896735832" watchObservedRunningTime="2025-12-04 13:17:07.62370232 +0000 UTC m=+5651.897998124" Dec 04 13:17:10 crc kubenswrapper[4979]: I1204 13:17:10.557593 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:17:10 crc kubenswrapper[4979]: I1204 13:17:10.558006 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:17:10 crc kubenswrapper[4979]: I1204 13:17:10.960765 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 13:17:14 crc kubenswrapper[4979]: I1204 13:17:14.970066 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 13:17:14 crc kubenswrapper[4979]: I1204 13:17:14.970512 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 13:17:15 crc kubenswrapper[4979]: I1204 13:17:15.557800 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 13:17:15 crc kubenswrapper[4979]: I1204 13:17:15.558187 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 13:17:15 crc kubenswrapper[4979]: I1204 13:17:15.960779 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 13:17:15 crc kubenswrapper[4979]: I1204 13:17:15.999161 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.052601 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.052640 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.546503 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ddrk2"] Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.549282 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.555621 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ddrk2"] Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.640681 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.641018 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.706474 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-catalog-content\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.706528 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-utilities\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.706952 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l62cm\" (UniqueName: \"kubernetes.io/projected/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-kube-api-access-l62cm\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.714185 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.808566 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-catalog-content\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.808615 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-utilities\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.808744 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l62cm\" (UniqueName: \"kubernetes.io/projected/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-kube-api-access-l62cm\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.809170 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-utilities\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.809494 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-catalog-content\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.837421 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l62cm\" (UniqueName: \"kubernetes.io/projected/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-kube-api-access-l62cm\") pod \"certified-operators-ddrk2\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:16 crc kubenswrapper[4979]: I1204 13:17:16.887084 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:17 crc kubenswrapper[4979]: I1204 13:17:17.534941 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ddrk2"] Dec 04 13:17:17 crc kubenswrapper[4979]: I1204 13:17:17.691193 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddrk2" event={"ID":"c86d1fa8-6661-463b-b9cd-f503e0f3bbab","Type":"ContainerStarted","Data":"e25281b230f695d93b5d09562f8dbc80b1d6d70b6475b4b684b0efda16c8cdb1"} Dec 04 13:17:18 crc kubenswrapper[4979]: I1204 13:17:18.704888 4979 generic.go:334] "Generic (PLEG): container finished" podID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerID="1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd" exitCode=0 Dec 04 13:17:18 crc kubenswrapper[4979]: I1204 13:17:18.704943 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddrk2" event={"ID":"c86d1fa8-6661-463b-b9cd-f503e0f3bbab","Type":"ContainerDied","Data":"1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd"} Dec 04 13:17:18 crc kubenswrapper[4979]: I1204 13:17:18.706709 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:17:19 crc kubenswrapper[4979]: I1204 13:17:19.714881 4979 generic.go:334] "Generic (PLEG): container finished" podID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerID="20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba" exitCode=0 Dec 04 13:17:19 crc kubenswrapper[4979]: I1204 13:17:19.714982 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddrk2" event={"ID":"c86d1fa8-6661-463b-b9cd-f503e0f3bbab","Type":"ContainerDied","Data":"20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba"} Dec 04 13:17:20 crc kubenswrapper[4979]: I1204 13:17:20.728755 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddrk2" event={"ID":"c86d1fa8-6661-463b-b9cd-f503e0f3bbab","Type":"ContainerStarted","Data":"5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72"} Dec 04 13:17:20 crc kubenswrapper[4979]: I1204 13:17:20.752480 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ddrk2" podStartSLOduration=3.343629859 podStartE2EDuration="4.752460171s" podCreationTimestamp="2025-12-04 13:17:16 +0000 UTC" firstStartedPulling="2025-12-04 13:17:18.70640669 +0000 UTC m=+5662.980702494" lastFinishedPulling="2025-12-04 13:17:20.115237002 +0000 UTC m=+5664.389532806" observedRunningTime="2025-12-04 13:17:20.745946429 +0000 UTC m=+5665.020242233" watchObservedRunningTime="2025-12-04 13:17:20.752460171 +0000 UTC m=+5665.026755965" Dec 04 13:17:24 crc kubenswrapper[4979]: I1204 13:17:24.974093 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 13:17:24 crc kubenswrapper[4979]: I1204 13:17:24.974711 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 13:17:24 crc kubenswrapper[4979]: I1204 13:17:24.975116 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 13:17:24 crc kubenswrapper[4979]: I1204 13:17:24.975170 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 13:17:24 crc kubenswrapper[4979]: I1204 13:17:24.979118 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 13:17:24 crc kubenswrapper[4979]: I1204 13:17:24.979894 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.185136 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc6fb8cf7-lz4np"] Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.186891 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.201756 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc6fb8cf7-lz4np"] Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.277011 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.278129 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-dns-svc\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.278424 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-config\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.278471 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/63c35ce1-6511-4d13-a9d3-5711ea936131-kube-api-access-d8pv2\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.278578 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.379838 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-config\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.379908 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/63c35ce1-6511-4d13-a9d3-5711ea936131-kube-api-access-d8pv2\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.379970 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.380027 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.380077 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-dns-svc\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.381220 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.381222 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-nb\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.381257 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-dns-svc\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.381925 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-config\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.408039 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/63c35ce1-6511-4d13-a9d3-5711ea936131-kube-api-access-d8pv2\") pod \"dnsmasq-dns-6bc6fb8cf7-lz4np\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.520346 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.564618 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.567104 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.569466 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 13:17:25 crc kubenswrapper[4979]: I1204 13:17:25.770510 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.010855 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc6fb8cf7-lz4np"] Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.777061 4979 generic.go:334] "Generic (PLEG): container finished" podID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerID="088a24958b64aee540902b0dd983b971c37705523281a2c000407c1d6c344b82" exitCode=0 Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.777158 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" event={"ID":"63c35ce1-6511-4d13-a9d3-5711ea936131","Type":"ContainerDied","Data":"088a24958b64aee540902b0dd983b971c37705523281a2c000407c1d6c344b82"} Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.777466 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" event={"ID":"63c35ce1-6511-4d13-a9d3-5711ea936131","Type":"ContainerStarted","Data":"6f599414d951a580d62d34f0787a787d9a10b3bfb1760790b711eadd2a7eddd0"} Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.887871 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.887939 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:26 crc kubenswrapper[4979]: I1204 13:17:26.939408 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:27 crc kubenswrapper[4979]: I1204 13:17:27.788969 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" event={"ID":"63c35ce1-6511-4d13-a9d3-5711ea936131","Type":"ContainerStarted","Data":"0e024b8b14d286ad83936afade740527d2d74aca76975ebe12ef25488aa19076"} Dec 04 13:17:27 crc kubenswrapper[4979]: I1204 13:17:27.789741 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:27 crc kubenswrapper[4979]: I1204 13:17:27.811546 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" podStartSLOduration=2.811530218 podStartE2EDuration="2.811530218s" podCreationTimestamp="2025-12-04 13:17:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:27.806907765 +0000 UTC m=+5672.081203589" watchObservedRunningTime="2025-12-04 13:17:27.811530218 +0000 UTC m=+5672.085826022" Dec 04 13:17:27 crc kubenswrapper[4979]: I1204 13:17:27.844481 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:27 crc kubenswrapper[4979]: I1204 13:17:27.901233 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ddrk2"] Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.041028 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.041092 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.041138 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.041896 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.041957 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" gracePeriod=600 Dec 04 13:17:28 crc kubenswrapper[4979]: E1204 13:17:28.166324 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.802651 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" exitCode=0 Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.803637 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005"} Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.803679 4979 scope.go:117] "RemoveContainer" containerID="f41cc400d512fbc84293fd7f01e9e77498ba3dd4183c76c64ed639782564147d" Dec 04 13:17:28 crc kubenswrapper[4979]: I1204 13:17:28.804561 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:17:28 crc kubenswrapper[4979]: E1204 13:17:28.804788 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:17:29 crc kubenswrapper[4979]: I1204 13:17:29.832860 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ddrk2" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="registry-server" containerID="cri-o://5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72" gracePeriod=2 Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.816231 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.858624 4979 generic.go:334] "Generic (PLEG): container finished" podID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerID="5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72" exitCode=0 Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.858696 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddrk2" event={"ID":"c86d1fa8-6661-463b-b9cd-f503e0f3bbab","Type":"ContainerDied","Data":"5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72"} Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.858749 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ddrk2" event={"ID":"c86d1fa8-6661-463b-b9cd-f503e0f3bbab","Type":"ContainerDied","Data":"e25281b230f695d93b5d09562f8dbc80b1d6d70b6475b4b684b0efda16c8cdb1"} Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.858754 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ddrk2" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.858775 4979 scope.go:117] "RemoveContainer" containerID="5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.893604 4979 scope.go:117] "RemoveContainer" containerID="20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.923509 4979 scope.go:117] "RemoveContainer" containerID="1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.975510 4979 scope.go:117] "RemoveContainer" containerID="5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72" Dec 04 13:17:30 crc kubenswrapper[4979]: E1204 13:17:30.976025 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72\": container with ID starting with 5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72 not found: ID does not exist" containerID="5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.976091 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72"} err="failed to get container status \"5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72\": rpc error: code = NotFound desc = could not find container \"5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72\": container with ID starting with 5f2a8e75ace884ecc39a62bb988e95ccc7e497bc802576fe596c153fc3df4c72 not found: ID does not exist" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.976126 4979 scope.go:117] "RemoveContainer" containerID="20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba" Dec 04 13:17:30 crc kubenswrapper[4979]: E1204 13:17:30.993606 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba\": container with ID starting with 20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba not found: ID does not exist" containerID="20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.993663 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba"} err="failed to get container status \"20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba\": rpc error: code = NotFound desc = could not find container \"20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba\": container with ID starting with 20076295695c737cc0aa3d5e101abd575c371d7bc5cb83f2bda3fbf9df1123ba not found: ID does not exist" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.993696 4979 scope.go:117] "RemoveContainer" containerID="1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.994059 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-utilities\") pod \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.994224 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-catalog-content\") pod \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " Dec 04 13:17:30 crc kubenswrapper[4979]: E1204 13:17:30.994224 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd\": container with ID starting with 1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd not found: ID does not exist" containerID="1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.994355 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd"} err="failed to get container status \"1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd\": rpc error: code = NotFound desc = could not find container \"1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd\": container with ID starting with 1582d11e0cae30b7103efcfb46baed222e2958d137a4341549ca2212bf5ce0bd not found: ID does not exist" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.994278 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l62cm\" (UniqueName: \"kubernetes.io/projected/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-kube-api-access-l62cm\") pod \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\" (UID: \"c86d1fa8-6661-463b-b9cd-f503e0f3bbab\") " Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.994826 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-utilities" (OuterVolumeSpecName: "utilities") pod "c86d1fa8-6661-463b-b9cd-f503e0f3bbab" (UID: "c86d1fa8-6661-463b-b9cd-f503e0f3bbab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:17:30 crc kubenswrapper[4979]: I1204 13:17:30.995324 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:31 crc kubenswrapper[4979]: I1204 13:17:31.001182 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-kube-api-access-l62cm" (OuterVolumeSpecName: "kube-api-access-l62cm") pod "c86d1fa8-6661-463b-b9cd-f503e0f3bbab" (UID: "c86d1fa8-6661-463b-b9cd-f503e0f3bbab"). InnerVolumeSpecName "kube-api-access-l62cm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:31 crc kubenswrapper[4979]: I1204 13:17:31.046546 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c86d1fa8-6661-463b-b9cd-f503e0f3bbab" (UID: "c86d1fa8-6661-463b-b9cd-f503e0f3bbab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:17:31 crc kubenswrapper[4979]: I1204 13:17:31.096612 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:31 crc kubenswrapper[4979]: I1204 13:17:31.096651 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l62cm\" (UniqueName: \"kubernetes.io/projected/c86d1fa8-6661-463b-b9cd-f503e0f3bbab-kube-api-access-l62cm\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:31 crc kubenswrapper[4979]: I1204 13:17:31.198701 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ddrk2"] Dec 04 13:17:31 crc kubenswrapper[4979]: I1204 13:17:31.207996 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ddrk2"] Dec 04 13:17:32 crc kubenswrapper[4979]: I1204 13:17:32.210322 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" path="/var/lib/kubelet/pods/c86d1fa8-6661-463b-b9cd-f503e0f3bbab/volumes" Dec 04 13:17:35 crc kubenswrapper[4979]: I1204 13:17:35.521483 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:17:35 crc kubenswrapper[4979]: I1204 13:17:35.602752 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d598bd765-9z7pb"] Dec 04 13:17:35 crc kubenswrapper[4979]: I1204 13:17:35.606959 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerName="dnsmasq-dns" containerID="cri-o://5a3ef93080fc9db8f226f566ab4397bc1ad4690bc98b9af793da0b10f60029a3" gracePeriod=10 Dec 04 13:17:35 crc kubenswrapper[4979]: I1204 13:17:35.908555 4979 generic.go:334] "Generic (PLEG): container finished" podID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerID="5a3ef93080fc9db8f226f566ab4397bc1ad4690bc98b9af793da0b10f60029a3" exitCode=0 Dec 04 13:17:35 crc kubenswrapper[4979]: I1204 13:17:35.908611 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" event={"ID":"ce29f304-4be2-4e51-8d84-c0d20815a70c","Type":"ContainerDied","Data":"5a3ef93080fc9db8f226f566ab4397bc1ad4690bc98b9af793da0b10f60029a3"} Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.111949 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.186868 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-nb\") pod \"ce29f304-4be2-4e51-8d84-c0d20815a70c\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.186969 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkrc8\" (UniqueName: \"kubernetes.io/projected/ce29f304-4be2-4e51-8d84-c0d20815a70c-kube-api-access-nkrc8\") pod \"ce29f304-4be2-4e51-8d84-c0d20815a70c\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.187024 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-config\") pod \"ce29f304-4be2-4e51-8d84-c0d20815a70c\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.187051 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-dns-svc\") pod \"ce29f304-4be2-4e51-8d84-c0d20815a70c\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.187082 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-sb\") pod \"ce29f304-4be2-4e51-8d84-c0d20815a70c\" (UID: \"ce29f304-4be2-4e51-8d84-c0d20815a70c\") " Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.197891 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce29f304-4be2-4e51-8d84-c0d20815a70c-kube-api-access-nkrc8" (OuterVolumeSpecName: "kube-api-access-nkrc8") pod "ce29f304-4be2-4e51-8d84-c0d20815a70c" (UID: "ce29f304-4be2-4e51-8d84-c0d20815a70c"). InnerVolumeSpecName "kube-api-access-nkrc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.235294 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ce29f304-4be2-4e51-8d84-c0d20815a70c" (UID: "ce29f304-4be2-4e51-8d84-c0d20815a70c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.244852 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-config" (OuterVolumeSpecName: "config") pod "ce29f304-4be2-4e51-8d84-c0d20815a70c" (UID: "ce29f304-4be2-4e51-8d84-c0d20815a70c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.249020 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ce29f304-4be2-4e51-8d84-c0d20815a70c" (UID: "ce29f304-4be2-4e51-8d84-c0d20815a70c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.253308 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ce29f304-4be2-4e51-8d84-c0d20815a70c" (UID: "ce29f304-4be2-4e51-8d84-c0d20815a70c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.290881 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.291509 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkrc8\" (UniqueName: \"kubernetes.io/projected/ce29f304-4be2-4e51-8d84-c0d20815a70c-kube-api-access-nkrc8\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.291529 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.291543 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.291579 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ce29f304-4be2-4e51-8d84-c0d20815a70c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.918894 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" event={"ID":"ce29f304-4be2-4e51-8d84-c0d20815a70c","Type":"ContainerDied","Data":"decc6b46a097403580ca08c689ef43d49902837f26dd1aa5428eb59ca57dd6fd"} Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.919033 4979 scope.go:117] "RemoveContainer" containerID="5a3ef93080fc9db8f226f566ab4397bc1ad4690bc98b9af793da0b10f60029a3" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.918985 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d598bd765-9z7pb" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.947267 4979 scope.go:117] "RemoveContainer" containerID="ef777ec80b28541c1ecf9d5f2631f33bcf851102d1fdd8e3b8736f4db605fd18" Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.969752 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d598bd765-9z7pb"] Dec 04 13:17:36 crc kubenswrapper[4979]: I1204 13:17:36.985268 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d598bd765-9z7pb"] Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.208978 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" path="/var/lib/kubelet/pods/ce29f304-4be2-4e51-8d84-c0d20815a70c/volumes" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.982812 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-b4xr9"] Dec 04 13:17:38 crc kubenswrapper[4979]: E1204 13:17:38.983179 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="registry-server" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983192 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="registry-server" Dec 04 13:17:38 crc kubenswrapper[4979]: E1204 13:17:38.983203 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="extract-utilities" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983209 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="extract-utilities" Dec 04 13:17:38 crc kubenswrapper[4979]: E1204 13:17:38.983228 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerName="init" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983234 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerName="init" Dec 04 13:17:38 crc kubenswrapper[4979]: E1204 13:17:38.983250 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="extract-content" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983255 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="extract-content" Dec 04 13:17:38 crc kubenswrapper[4979]: E1204 13:17:38.983267 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerName="dnsmasq-dns" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983273 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerName="dnsmasq-dns" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983457 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce29f304-4be2-4e51-8d84-c0d20815a70c" containerName="dnsmasq-dns" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.983470 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c86d1fa8-6661-463b-b9cd-f503e0f3bbab" containerName="registry-server" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.984061 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.997663 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8f96-account-create-update-pcds5"] Dec 04 13:17:38 crc kubenswrapper[4979]: I1204 13:17:38.999209 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.001086 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.013465 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b4xr9"] Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.028867 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8f96-account-create-update-pcds5"] Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.041319 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnc76\" (UniqueName: \"kubernetes.io/projected/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-kube-api-access-nnc76\") pod \"cinder-db-create-b4xr9\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.041650 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-operator-scripts\") pod \"cinder-db-create-b4xr9\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.143121 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnc76\" (UniqueName: \"kubernetes.io/projected/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-kube-api-access-nnc76\") pod \"cinder-db-create-b4xr9\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.143177 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f7be7de-5236-490b-b001-a9bfbb6f4575-operator-scripts\") pod \"cinder-8f96-account-create-update-pcds5\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.143199 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfb9k\" (UniqueName: \"kubernetes.io/projected/5f7be7de-5236-490b-b001-a9bfbb6f4575-kube-api-access-dfb9k\") pod \"cinder-8f96-account-create-update-pcds5\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.143254 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-operator-scripts\") pod \"cinder-db-create-b4xr9\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.143982 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-operator-scripts\") pod \"cinder-db-create-b4xr9\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.162667 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnc76\" (UniqueName: \"kubernetes.io/projected/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-kube-api-access-nnc76\") pod \"cinder-db-create-b4xr9\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.245596 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f7be7de-5236-490b-b001-a9bfbb6f4575-operator-scripts\") pod \"cinder-8f96-account-create-update-pcds5\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.245899 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfb9k\" (UniqueName: \"kubernetes.io/projected/5f7be7de-5236-490b-b001-a9bfbb6f4575-kube-api-access-dfb9k\") pod \"cinder-8f96-account-create-update-pcds5\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.246405 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f7be7de-5236-490b-b001-a9bfbb6f4575-operator-scripts\") pod \"cinder-8f96-account-create-update-pcds5\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.261820 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfb9k\" (UniqueName: \"kubernetes.io/projected/5f7be7de-5236-490b-b001-a9bfbb6f4575-kube-api-access-dfb9k\") pod \"cinder-8f96-account-create-update-pcds5\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.307002 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.328790 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.748587 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b4xr9"] Dec 04 13:17:39 crc kubenswrapper[4979]: W1204 13:17:39.749753 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d8f7d50_afdf_4be5_b4ce_d7cf262369f2.slice/crio-da7a69ae82ada7eafcc1fe516b832e79b80cb6579662079ae1b359232e74feaa WatchSource:0}: Error finding container da7a69ae82ada7eafcc1fe516b832e79b80cb6579662079ae1b359232e74feaa: Status 404 returned error can't find the container with id da7a69ae82ada7eafcc1fe516b832e79b80cb6579662079ae1b359232e74feaa Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.809892 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8f96-account-create-update-pcds5"] Dec 04 13:17:39 crc kubenswrapper[4979]: W1204 13:17:39.813896 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5f7be7de_5236_490b_b001_a9bfbb6f4575.slice/crio-4ee62587827e7e0bd3fd60ce9140e2c9d87d48d11492928b5e61bd6157feb413 WatchSource:0}: Error finding container 4ee62587827e7e0bd3fd60ce9140e2c9d87d48d11492928b5e61bd6157feb413: Status 404 returned error can't find the container with id 4ee62587827e7e0bd3fd60ce9140e2c9d87d48d11492928b5e61bd6157feb413 Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.950588 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f96-account-create-update-pcds5" event={"ID":"5f7be7de-5236-490b-b001-a9bfbb6f4575","Type":"ContainerStarted","Data":"f95b5dff05ff1aa32bd72dd2810c3ca8678f54b7f48acace75412084cb40fe5a"} Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.950633 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f96-account-create-update-pcds5" event={"ID":"5f7be7de-5236-490b-b001-a9bfbb6f4575","Type":"ContainerStarted","Data":"4ee62587827e7e0bd3fd60ce9140e2c9d87d48d11492928b5e61bd6157feb413"} Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.952873 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4xr9" event={"ID":"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2","Type":"ContainerStarted","Data":"77be6ee27918f9b9d877e3fd44caeb0aec67223551070665f7aa95b9636e1b7d"} Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.952911 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4xr9" event={"ID":"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2","Type":"ContainerStarted","Data":"da7a69ae82ada7eafcc1fe516b832e79b80cb6579662079ae1b359232e74feaa"} Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.966512 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-8f96-account-create-update-pcds5" podStartSLOduration=1.966494011 podStartE2EDuration="1.966494011s" podCreationTimestamp="2025-12-04 13:17:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:39.965547386 +0000 UTC m=+5684.239843190" watchObservedRunningTime="2025-12-04 13:17:39.966494011 +0000 UTC m=+5684.240789815" Dec 04 13:17:39 crc kubenswrapper[4979]: I1204 13:17:39.990700 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-b4xr9" podStartSLOduration=1.990681544 podStartE2EDuration="1.990681544s" podCreationTimestamp="2025-12-04 13:17:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:39.980382921 +0000 UTC m=+5684.254678725" watchObservedRunningTime="2025-12-04 13:17:39.990681544 +0000 UTC m=+5684.264977348" Dec 04 13:17:40 crc kubenswrapper[4979]: I1204 13:17:40.963208 4979 generic.go:334] "Generic (PLEG): container finished" podID="5f7be7de-5236-490b-b001-a9bfbb6f4575" containerID="f95b5dff05ff1aa32bd72dd2810c3ca8678f54b7f48acace75412084cb40fe5a" exitCode=0 Dec 04 13:17:40 crc kubenswrapper[4979]: I1204 13:17:40.963281 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f96-account-create-update-pcds5" event={"ID":"5f7be7de-5236-490b-b001-a9bfbb6f4575","Type":"ContainerDied","Data":"f95b5dff05ff1aa32bd72dd2810c3ca8678f54b7f48acace75412084cb40fe5a"} Dec 04 13:17:40 crc kubenswrapper[4979]: I1204 13:17:40.972391 4979 generic.go:334] "Generic (PLEG): container finished" podID="0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" containerID="77be6ee27918f9b9d877e3fd44caeb0aec67223551070665f7aa95b9636e1b7d" exitCode=0 Dec 04 13:17:40 crc kubenswrapper[4979]: I1204 13:17:40.973035 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4xr9" event={"ID":"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2","Type":"ContainerDied","Data":"77be6ee27918f9b9d877e3fd44caeb0aec67223551070665f7aa95b9636e1b7d"} Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.469963 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.480388 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.599705 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-operator-scripts\") pod \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.599779 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnc76\" (UniqueName: \"kubernetes.io/projected/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-kube-api-access-nnc76\") pod \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\" (UID: \"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2\") " Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.599818 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f7be7de-5236-490b-b001-a9bfbb6f4575-operator-scripts\") pod \"5f7be7de-5236-490b-b001-a9bfbb6f4575\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.599950 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfb9k\" (UniqueName: \"kubernetes.io/projected/5f7be7de-5236-490b-b001-a9bfbb6f4575-kube-api-access-dfb9k\") pod \"5f7be7de-5236-490b-b001-a9bfbb6f4575\" (UID: \"5f7be7de-5236-490b-b001-a9bfbb6f4575\") " Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.600542 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" (UID: "0d8f7d50-afdf-4be5-b4ce-d7cf262369f2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.600542 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f7be7de-5236-490b-b001-a9bfbb6f4575-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f7be7de-5236-490b-b001-a9bfbb6f4575" (UID: "5f7be7de-5236-490b-b001-a9bfbb6f4575"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.606248 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f7be7de-5236-490b-b001-a9bfbb6f4575-kube-api-access-dfb9k" (OuterVolumeSpecName: "kube-api-access-dfb9k") pod "5f7be7de-5236-490b-b001-a9bfbb6f4575" (UID: "5f7be7de-5236-490b-b001-a9bfbb6f4575"). InnerVolumeSpecName "kube-api-access-dfb9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.606726 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-kube-api-access-nnc76" (OuterVolumeSpecName: "kube-api-access-nnc76") pod "0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" (UID: "0d8f7d50-afdf-4be5-b4ce-d7cf262369f2"). InnerVolumeSpecName "kube-api-access-nnc76". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.701805 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.701848 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnc76\" (UniqueName: \"kubernetes.io/projected/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2-kube-api-access-nnc76\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.701867 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f7be7de-5236-490b-b001-a9bfbb6f4575-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.701879 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfb9k\" (UniqueName: \"kubernetes.io/projected/5f7be7de-5236-490b-b001-a9bfbb6f4575-kube-api-access-dfb9k\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.990647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b4xr9" event={"ID":"0d8f7d50-afdf-4be5-b4ce-d7cf262369f2","Type":"ContainerDied","Data":"da7a69ae82ada7eafcc1fe516b832e79b80cb6579662079ae1b359232e74feaa"} Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.990704 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da7a69ae82ada7eafcc1fe516b832e79b80cb6579662079ae1b359232e74feaa" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.990659 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b4xr9" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.993553 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8f96-account-create-update-pcds5" event={"ID":"5f7be7de-5236-490b-b001-a9bfbb6f4575","Type":"ContainerDied","Data":"4ee62587827e7e0bd3fd60ce9140e2c9d87d48d11492928b5e61bd6157feb413"} Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.993589 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8f96-account-create-update-pcds5" Dec 04 13:17:42 crc kubenswrapper[4979]: I1204 13:17:42.993599 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ee62587827e7e0bd3fd60ce9140e2c9d87d48d11492928b5e61bd6157feb413" Dec 04 13:17:43 crc kubenswrapper[4979]: I1204 13:17:43.199882 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:17:43 crc kubenswrapper[4979]: E1204 13:17:43.200430 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.330146 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-55dkm"] Dec 04 13:17:44 crc kubenswrapper[4979]: E1204 13:17:44.330873 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" containerName="mariadb-database-create" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.330888 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" containerName="mariadb-database-create" Dec 04 13:17:44 crc kubenswrapper[4979]: E1204 13:17:44.330936 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f7be7de-5236-490b-b001-a9bfbb6f4575" containerName="mariadb-account-create-update" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.330946 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f7be7de-5236-490b-b001-a9bfbb6f4575" containerName="mariadb-account-create-update" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.331146 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" containerName="mariadb-database-create" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.331169 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f7be7de-5236-490b-b001-a9bfbb6f4575" containerName="mariadb-account-create-update" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.331869 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.339815 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.340008 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9mrw8" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.340538 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.347795 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-55dkm"] Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.432076 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-db-sync-config-data\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.432829 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-combined-ca-bundle\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.432898 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-scripts\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.433150 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-config-data\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.433193 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-etc-machine-id\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.433252 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s47zw\" (UniqueName: \"kubernetes.io/projected/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-kube-api-access-s47zw\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.535735 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-combined-ca-bundle\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.536216 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-scripts\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.536414 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-config-data\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.536561 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-etc-machine-id\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.536664 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-etc-machine-id\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.536684 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s47zw\" (UniqueName: \"kubernetes.io/projected/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-kube-api-access-s47zw\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.536789 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-db-sync-config-data\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.543217 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-scripts\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.543348 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-combined-ca-bundle\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.543639 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-db-sync-config-data\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.543899 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-config-data\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.557992 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s47zw\" (UniqueName: \"kubernetes.io/projected/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-kube-api-access-s47zw\") pod \"cinder-db-sync-55dkm\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:44 crc kubenswrapper[4979]: I1204 13:17:44.650680 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:45 crc kubenswrapper[4979]: I1204 13:17:45.112369 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-55dkm"] Dec 04 13:17:46 crc kubenswrapper[4979]: I1204 13:17:46.022094 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-55dkm" event={"ID":"31eb23cc-ca38-4fa6-926b-549b3a4eb35a","Type":"ContainerStarted","Data":"9dea7d5bc9a412bd09350b4cbeed6fdffea9b2665cfc71a3002c95e093136be8"} Dec 04 13:17:46 crc kubenswrapper[4979]: I1204 13:17:46.022467 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-55dkm" event={"ID":"31eb23cc-ca38-4fa6-926b-549b3a4eb35a","Type":"ContainerStarted","Data":"d7b93f43b34b890387935ca1e096b517be36613f1e8a1c153f80d8d9239adb33"} Dec 04 13:17:46 crc kubenswrapper[4979]: I1204 13:17:46.038274 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-55dkm" podStartSLOduration=2.038257312 podStartE2EDuration="2.038257312s" podCreationTimestamp="2025-12-04 13:17:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:46.037776688 +0000 UTC m=+5690.312072492" watchObservedRunningTime="2025-12-04 13:17:46.038257312 +0000 UTC m=+5690.312553116" Dec 04 13:17:49 crc kubenswrapper[4979]: I1204 13:17:49.047042 4979 generic.go:334] "Generic (PLEG): container finished" podID="31eb23cc-ca38-4fa6-926b-549b3a4eb35a" containerID="9dea7d5bc9a412bd09350b4cbeed6fdffea9b2665cfc71a3002c95e093136be8" exitCode=0 Dec 04 13:17:49 crc kubenswrapper[4979]: I1204 13:17:49.047098 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-55dkm" event={"ID":"31eb23cc-ca38-4fa6-926b-549b3a4eb35a","Type":"ContainerDied","Data":"9dea7d5bc9a412bd09350b4cbeed6fdffea9b2665cfc71a3002c95e093136be8"} Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.376252 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.455734 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s47zw\" (UniqueName: \"kubernetes.io/projected/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-kube-api-access-s47zw\") pod \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.455793 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-scripts\") pod \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.455843 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-db-sync-config-data\") pod \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.455989 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-combined-ca-bundle\") pod \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.456080 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-config-data\") pod \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.456110 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-etc-machine-id\") pod \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\" (UID: \"31eb23cc-ca38-4fa6-926b-549b3a4eb35a\") " Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.456347 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "31eb23cc-ca38-4fa6-926b-549b3a4eb35a" (UID: "31eb23cc-ca38-4fa6-926b-549b3a4eb35a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.456757 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.461402 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "31eb23cc-ca38-4fa6-926b-549b3a4eb35a" (UID: "31eb23cc-ca38-4fa6-926b-549b3a4eb35a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.461720 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-scripts" (OuterVolumeSpecName: "scripts") pod "31eb23cc-ca38-4fa6-926b-549b3a4eb35a" (UID: "31eb23cc-ca38-4fa6-926b-549b3a4eb35a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.461935 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-kube-api-access-s47zw" (OuterVolumeSpecName: "kube-api-access-s47zw") pod "31eb23cc-ca38-4fa6-926b-549b3a4eb35a" (UID: "31eb23cc-ca38-4fa6-926b-549b3a4eb35a"). InnerVolumeSpecName "kube-api-access-s47zw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.484315 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31eb23cc-ca38-4fa6-926b-549b3a4eb35a" (UID: "31eb23cc-ca38-4fa6-926b-549b3a4eb35a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.504166 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-config-data" (OuterVolumeSpecName: "config-data") pod "31eb23cc-ca38-4fa6-926b-549b3a4eb35a" (UID: "31eb23cc-ca38-4fa6-926b-549b3a4eb35a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.558043 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s47zw\" (UniqueName: \"kubernetes.io/projected/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-kube-api-access-s47zw\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.558082 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.558092 4979 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.558100 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:50 crc kubenswrapper[4979]: I1204 13:17:50.558108 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31eb23cc-ca38-4fa6-926b-549b3a4eb35a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.071014 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-55dkm" event={"ID":"31eb23cc-ca38-4fa6-926b-549b3a4eb35a","Type":"ContainerDied","Data":"d7b93f43b34b890387935ca1e096b517be36613f1e8a1c153f80d8d9239adb33"} Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.071627 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7b93f43b34b890387935ca1e096b517be36613f1e8a1c153f80d8d9239adb33" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.071444 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-55dkm" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.439381 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9fb795d9-q7m4c"] Dec 04 13:17:51 crc kubenswrapper[4979]: E1204 13:17:51.439922 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31eb23cc-ca38-4fa6-926b-549b3a4eb35a" containerName="cinder-db-sync" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.439939 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="31eb23cc-ca38-4fa6-926b-549b3a4eb35a" containerName="cinder-db-sync" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.440185 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="31eb23cc-ca38-4fa6-926b-549b3a4eb35a" containerName="cinder-db-sync" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.441433 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.464963 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9fb795d9-q7m4c"] Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.474031 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnwgn\" (UniqueName: \"kubernetes.io/projected/3bb84134-7a17-451e-bf57-b70469ec7581-kube-api-access-fnwgn\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.474194 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-sb\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.474249 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.474321 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-nb\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.474349 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-config\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.536971 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.538987 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.551893 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.553580 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.553881 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.553897 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.554012 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-9mrw8" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578544 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt9j9\" (UniqueName: \"kubernetes.io/projected/dbdd03c8-bb67-4284-bc45-be2401c5d665-kube-api-access-wt9j9\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578606 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578648 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data-custom\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578677 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-sb\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578695 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbdd03c8-bb67-4284-bc45-be2401c5d665-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578732 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-scripts\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578756 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578812 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-nb\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578837 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-config\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578888 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnwgn\" (UniqueName: \"kubernetes.io/projected/3bb84134-7a17-451e-bf57-b70469ec7581-kube-api-access-fnwgn\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578918 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbdd03c8-bb67-4284-bc45-be2401c5d665-logs\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.578936 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.580460 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-config\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.581418 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-sb\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.581620 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.583449 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-nb\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.609464 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnwgn\" (UniqueName: \"kubernetes.io/projected/3bb84134-7a17-451e-bf57-b70469ec7581-kube-api-access-fnwgn\") pod \"dnsmasq-dns-9fb795d9-q7m4c\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.680907 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbdd03c8-bb67-4284-bc45-be2401c5d665-logs\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.680964 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681061 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt9j9\" (UniqueName: \"kubernetes.io/projected/dbdd03c8-bb67-4284-bc45-be2401c5d665-kube-api-access-wt9j9\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681081 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681107 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data-custom\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681133 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbdd03c8-bb67-4284-bc45-be2401c5d665-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681156 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-scripts\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681361 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbdd03c8-bb67-4284-bc45-be2401c5d665-logs\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.681840 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbdd03c8-bb67-4284-bc45-be2401c5d665-etc-machine-id\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.684498 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-scripts\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.684910 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data-custom\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.684979 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.685207 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.713120 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt9j9\" (UniqueName: \"kubernetes.io/projected/dbdd03c8-bb67-4284-bc45-be2401c5d665-kube-api-access-wt9j9\") pod \"cinder-api-0\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " pod="openstack/cinder-api-0" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.776072 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:51 crc kubenswrapper[4979]: I1204 13:17:51.867041 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 13:17:52 crc kubenswrapper[4979]: I1204 13:17:52.275791 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9fb795d9-q7m4c"] Dec 04 13:17:52 crc kubenswrapper[4979]: I1204 13:17:52.426819 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:17:52 crc kubenswrapper[4979]: W1204 13:17:52.432292 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbdd03c8_bb67_4284_bc45_be2401c5d665.slice/crio-ef5de585b1c540676a20bbc2cc2f5cdbc6d49ffb9b1a13facf2f5a638e5d4eff WatchSource:0}: Error finding container ef5de585b1c540676a20bbc2cc2f5cdbc6d49ffb9b1a13facf2f5a638e5d4eff: Status 404 returned error can't find the container with id ef5de585b1c540676a20bbc2cc2f5cdbc6d49ffb9b1a13facf2f5a638e5d4eff Dec 04 13:17:53 crc kubenswrapper[4979]: I1204 13:17:53.118225 4979 generic.go:334] "Generic (PLEG): container finished" podID="3bb84134-7a17-451e-bf57-b70469ec7581" containerID="dd404aafd97ae5cca6a005f48098e545faada82d16b9c4b5cd05606914e07047" exitCode=0 Dec 04 13:17:53 crc kubenswrapper[4979]: I1204 13:17:53.118752 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" event={"ID":"3bb84134-7a17-451e-bf57-b70469ec7581","Type":"ContainerDied","Data":"dd404aafd97ae5cca6a005f48098e545faada82d16b9c4b5cd05606914e07047"} Dec 04 13:17:53 crc kubenswrapper[4979]: I1204 13:17:53.118784 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" event={"ID":"3bb84134-7a17-451e-bf57-b70469ec7581","Type":"ContainerStarted","Data":"814050b30a584ca25b99647b5daed0e88dc231f6c973012262b51e48fc9a2b71"} Dec 04 13:17:53 crc kubenswrapper[4979]: I1204 13:17:53.123777 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbdd03c8-bb67-4284-bc45-be2401c5d665","Type":"ContainerStarted","Data":"f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1"} Dec 04 13:17:53 crc kubenswrapper[4979]: I1204 13:17:53.123818 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbdd03c8-bb67-4284-bc45-be2401c5d665","Type":"ContainerStarted","Data":"ef5de585b1c540676a20bbc2cc2f5cdbc6d49ffb9b1a13facf2f5a638e5d4eff"} Dec 04 13:17:54 crc kubenswrapper[4979]: I1204 13:17:54.135046 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbdd03c8-bb67-4284-bc45-be2401c5d665","Type":"ContainerStarted","Data":"ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1"} Dec 04 13:17:54 crc kubenswrapper[4979]: I1204 13:17:54.135559 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 13:17:54 crc kubenswrapper[4979]: I1204 13:17:54.138081 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" event={"ID":"3bb84134-7a17-451e-bf57-b70469ec7581","Type":"ContainerStarted","Data":"f040e9cda831fce6c18d75bbcf1a459850e3ec2e521cbc11db2f22899afc862f"} Dec 04 13:17:54 crc kubenswrapper[4979]: I1204 13:17:54.138309 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:17:54 crc kubenswrapper[4979]: I1204 13:17:54.152704 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.152682803 podStartE2EDuration="3.152682803s" podCreationTimestamp="2025-12-04 13:17:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:54.148560753 +0000 UTC m=+5698.422856557" watchObservedRunningTime="2025-12-04 13:17:54.152682803 +0000 UTC m=+5698.426978607" Dec 04 13:17:54 crc kubenswrapper[4979]: I1204 13:17:54.175817 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" podStartSLOduration=3.175796247 podStartE2EDuration="3.175796247s" podCreationTimestamp="2025-12-04 13:17:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:17:54.165820601 +0000 UTC m=+5698.440116405" watchObservedRunningTime="2025-12-04 13:17:54.175796247 +0000 UTC m=+5698.450092051" Dec 04 13:17:58 crc kubenswrapper[4979]: I1204 13:17:58.200079 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:17:58 crc kubenswrapper[4979]: E1204 13:17:58.201514 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:18:01 crc kubenswrapper[4979]: I1204 13:18:01.777471 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:18:01 crc kubenswrapper[4979]: I1204 13:18:01.855318 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc6fb8cf7-lz4np"] Dec 04 13:18:01 crc kubenswrapper[4979]: I1204 13:18:01.855758 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerName="dnsmasq-dns" containerID="cri-o://0e024b8b14d286ad83936afade740527d2d74aca76975ebe12ef25488aa19076" gracePeriod=10 Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.223642 4979 generic.go:334] "Generic (PLEG): container finished" podID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerID="0e024b8b14d286ad83936afade740527d2d74aca76975ebe12ef25488aa19076" exitCode=0 Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.227437 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" event={"ID":"63c35ce1-6511-4d13-a9d3-5711ea936131","Type":"ContainerDied","Data":"0e024b8b14d286ad83936afade740527d2d74aca76975ebe12ef25488aa19076"} Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.503586 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.700482 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-dns-svc\") pod \"63c35ce1-6511-4d13-a9d3-5711ea936131\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.700734 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-sb\") pod \"63c35ce1-6511-4d13-a9d3-5711ea936131\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.700791 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-nb\") pod \"63c35ce1-6511-4d13-a9d3-5711ea936131\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.700845 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-config\") pod \"63c35ce1-6511-4d13-a9d3-5711ea936131\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.700882 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/63c35ce1-6511-4d13-a9d3-5711ea936131-kube-api-access-d8pv2\") pod \"63c35ce1-6511-4d13-a9d3-5711ea936131\" (UID: \"63c35ce1-6511-4d13-a9d3-5711ea936131\") " Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.722277 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63c35ce1-6511-4d13-a9d3-5711ea936131-kube-api-access-d8pv2" (OuterVolumeSpecName: "kube-api-access-d8pv2") pod "63c35ce1-6511-4d13-a9d3-5711ea936131" (UID: "63c35ce1-6511-4d13-a9d3-5711ea936131"). InnerVolumeSpecName "kube-api-access-d8pv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.763123 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "63c35ce1-6511-4d13-a9d3-5711ea936131" (UID: "63c35ce1-6511-4d13-a9d3-5711ea936131"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.768957 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-config" (OuterVolumeSpecName: "config") pod "63c35ce1-6511-4d13-a9d3-5711ea936131" (UID: "63c35ce1-6511-4d13-a9d3-5711ea936131"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.775885 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "63c35ce1-6511-4d13-a9d3-5711ea936131" (UID: "63c35ce1-6511-4d13-a9d3-5711ea936131"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.781214 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "63c35ce1-6511-4d13-a9d3-5711ea936131" (UID: "63c35ce1-6511-4d13-a9d3-5711ea936131"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.802734 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.802993 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.803085 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.803159 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8pv2\" (UniqueName: \"kubernetes.io/projected/63c35ce1-6511-4d13-a9d3-5711ea936131-kube-api-access-d8pv2\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:02 crc kubenswrapper[4979]: I1204 13:18:02.803225 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/63c35ce1-6511-4d13-a9d3-5711ea936131-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.235172 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" event={"ID":"63c35ce1-6511-4d13-a9d3-5711ea936131","Type":"ContainerDied","Data":"6f599414d951a580d62d34f0787a787d9a10b3bfb1760790b711eadd2a7eddd0"} Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.235240 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc6fb8cf7-lz4np" Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.235591 4979 scope.go:117] "RemoveContainer" containerID="0e024b8b14d286ad83936afade740527d2d74aca76975ebe12ef25488aa19076" Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.275352 4979 scope.go:117] "RemoveContainer" containerID="088a24958b64aee540902b0dd983b971c37705523281a2c000407c1d6c344b82" Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.302066 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc6fb8cf7-lz4np"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.317076 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc6fb8cf7-lz4np"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.648684 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.648892 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="83299115-b631-4e6e-901a-e82944e5c21c" containerName="nova-cell0-conductor-conductor" containerID="cri-o://07bbe03959b56919a54c2cc12b07d0442d8ce804b250d0a6a5d7f7d887d3d15a" gracePeriod=30 Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.658455 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.659144 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" containerName="nova-scheduler-scheduler" containerID="cri-o://4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d" gracePeriod=30 Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.671109 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.671442 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-log" containerID="cri-o://1da2e815fda862ad0ddafcd951202e197e5c13bd9f123d32d3ec9bc4c29e8edd" gracePeriod=30 Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.671533 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-metadata" containerID="cri-o://4c9a940c9aefdc19229ce43bd16b871d1dbc397f9bbd271044a98121efd9b57c" gracePeriod=30 Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.685947 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.686224 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://cb835efb2f8c182dc2aa148b9e0c0f19d2f55e14f952ef358c8d134fe28d9968" gracePeriod=30 Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.697547 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.697884 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-log" containerID="cri-o://6aca61a38fb477a9267cb4b0138b2346e65ac917289490a6259f0aff4c6f5466" gracePeriod=30 Dec 04 13:18:03 crc kubenswrapper[4979]: I1204 13:18:03.697967 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-api" containerID="cri-o://8e6430c1953b1184982391fbe8caa201a15a1e05432d4e5e5b07895d2aa2f71d" gracePeriod=30 Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.148827 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.227543 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" path="/var/lib/kubelet/pods/63c35ce1-6511-4d13-a9d3-5711ea936131/volumes" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.297249 4979 generic.go:334] "Generic (PLEG): container finished" podID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerID="1da2e815fda862ad0ddafcd951202e197e5c13bd9f123d32d3ec9bc4c29e8edd" exitCode=143 Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.297398 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bda865f5-85ba-4d30-84e7-057bc6e2e1c9","Type":"ContainerDied","Data":"1da2e815fda862ad0ddafcd951202e197e5c13bd9f123d32d3ec9bc4c29e8edd"} Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.301464 4979 generic.go:334] "Generic (PLEG): container finished" podID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" containerID="cb835efb2f8c182dc2aa148b9e0c0f19d2f55e14f952ef358c8d134fe28d9968" exitCode=0 Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.301555 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"249ad63a-955e-4fe2-b30a-4534dd49f8ef","Type":"ContainerDied","Data":"cb835efb2f8c182dc2aa148b9e0c0f19d2f55e14f952ef358c8d134fe28d9968"} Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.305701 4979 generic.go:334] "Generic (PLEG): container finished" podID="e28a7327-643f-45f1-848c-d971f083136a" containerID="6aca61a38fb477a9267cb4b0138b2346e65ac917289490a6259f0aff4c6f5466" exitCode=143 Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.305748 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e28a7327-643f-45f1-848c-d971f083136a","Type":"ContainerDied","Data":"6aca61a38fb477a9267cb4b0138b2346e65ac917289490a6259f0aff4c6f5466"} Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.657398 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.840705 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-config-data\") pod \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.841143 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-combined-ca-bundle\") pod \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.841274 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bcvxk\" (UniqueName: \"kubernetes.io/projected/249ad63a-955e-4fe2-b30a-4534dd49f8ef-kube-api-access-bcvxk\") pod \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\" (UID: \"249ad63a-955e-4fe2-b30a-4534dd49f8ef\") " Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.851112 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/249ad63a-955e-4fe2-b30a-4534dd49f8ef-kube-api-access-bcvxk" (OuterVolumeSpecName: "kube-api-access-bcvxk") pod "249ad63a-955e-4fe2-b30a-4534dd49f8ef" (UID: "249ad63a-955e-4fe2-b30a-4534dd49f8ef"). InnerVolumeSpecName "kube-api-access-bcvxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.873043 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "249ad63a-955e-4fe2-b30a-4534dd49f8ef" (UID: "249ad63a-955e-4fe2-b30a-4534dd49f8ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.879320 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-config-data" (OuterVolumeSpecName: "config-data") pod "249ad63a-955e-4fe2-b30a-4534dd49f8ef" (UID: "249ad63a-955e-4fe2-b30a-4534dd49f8ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.944883 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.944926 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/249ad63a-955e-4fe2-b30a-4534dd49f8ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:04 crc kubenswrapper[4979]: I1204 13:18:04.944942 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bcvxk\" (UniqueName: \"kubernetes.io/projected/249ad63a-955e-4fe2-b30a-4534dd49f8ef-kube-api-access-bcvxk\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.034838 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.150025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nblch\" (UniqueName: \"kubernetes.io/projected/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-kube-api-access-nblch\") pod \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.150130 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-config-data\") pod \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.150252 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-combined-ca-bundle\") pod \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\" (UID: \"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5\") " Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.155958 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-kube-api-access-nblch" (OuterVolumeSpecName: "kube-api-access-nblch") pod "665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" (UID: "665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5"). InnerVolumeSpecName "kube-api-access-nblch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.177015 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-config-data" (OuterVolumeSpecName: "config-data") pod "665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" (UID: "665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.180237 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" (UID: "665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.251982 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.252028 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.252044 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nblch\" (UniqueName: \"kubernetes.io/projected/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5-kube-api-access-nblch\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.321657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"249ad63a-955e-4fe2-b30a-4534dd49f8ef","Type":"ContainerDied","Data":"903a00911c3dcf16f29a681e511d91d4cbd4e206f29321c2068e402004b339df"} Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.321719 4979 scope.go:117] "RemoveContainer" containerID="cb835efb2f8c182dc2aa148b9e0c0f19d2f55e14f952ef358c8d134fe28d9968" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.321888 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.331778 4979 generic.go:334] "Generic (PLEG): container finished" podID="665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" containerID="4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d" exitCode=0 Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.331829 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5","Type":"ContainerDied","Data":"4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d"} Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.331858 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5","Type":"ContainerDied","Data":"d3002c585acdc293a4cbe1f144e8c0e91190cfe1c3f3cb0e5c24723ae8f7a0a1"} Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.331911 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.367399 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.369405 4979 scope.go:117] "RemoveContainer" containerID="4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.380687 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.390876 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.399147 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.404325 4979 scope.go:117] "RemoveContainer" containerID="4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d" Dec 04 13:18:05 crc kubenswrapper[4979]: E1204 13:18:05.404713 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d\": container with ID starting with 4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d not found: ID does not exist" containerID="4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.404754 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d"} err="failed to get container status \"4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d\": rpc error: code = NotFound desc = could not find container \"4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d\": container with ID starting with 4ab5b70ab8ffa7f36e0c693c51213bc0dcfb5eab47f4d0ae31d5bf13f835105d not found: ID does not exist" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.411365 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: E1204 13:18:05.425582 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.425631 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 13:18:05 crc kubenswrapper[4979]: E1204 13:18:05.425646 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerName="dnsmasq-dns" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.425655 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerName="dnsmasq-dns" Dec 04 13:18:05 crc kubenswrapper[4979]: E1204 13:18:05.425688 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerName="init" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.425696 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerName="init" Dec 04 13:18:05 crc kubenswrapper[4979]: E1204 13:18:05.425726 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" containerName="nova-scheduler-scheduler" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.425735 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" containerName="nova-scheduler-scheduler" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.426015 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" containerName="nova-cell1-novncproxy-novncproxy" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.426044 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="63c35ce1-6511-4d13-a9d3-5711ea936131" containerName="dnsmasq-dns" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.426062 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" containerName="nova-scheduler-scheduler" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.426789 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.427470 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.431250 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.439940 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.440044 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.450374 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.481398 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.564397 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgkx5\" (UniqueName: \"kubernetes.io/projected/b0826aec-f059-4463-9942-3c84bd8a556d-kube-api-access-fgkx5\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.564480 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0826aec-f059-4463-9942-3c84bd8a556d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.564523 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-config-data\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.564558 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0826aec-f059-4463-9942-3c84bd8a556d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.564663 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqzq8\" (UniqueName: \"kubernetes.io/projected/cf1145a5-ea36-490c-9360-98feed19a052-kube-api-access-zqzq8\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.564704 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.669400 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqzq8\" (UniqueName: \"kubernetes.io/projected/cf1145a5-ea36-490c-9360-98feed19a052-kube-api-access-zqzq8\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.669489 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.669523 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgkx5\" (UniqueName: \"kubernetes.io/projected/b0826aec-f059-4463-9942-3c84bd8a556d-kube-api-access-fgkx5\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.669575 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0826aec-f059-4463-9942-3c84bd8a556d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.669611 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-config-data\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.669649 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0826aec-f059-4463-9942-3c84bd8a556d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.691167 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-config-data\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.696497 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0826aec-f059-4463-9942-3c84bd8a556d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.696950 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.699506 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqzq8\" (UniqueName: \"kubernetes.io/projected/cf1145a5-ea36-490c-9360-98feed19a052-kube-api-access-zqzq8\") pod \"nova-scheduler-0\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " pod="openstack/nova-scheduler-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.703661 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgkx5\" (UniqueName: \"kubernetes.io/projected/b0826aec-f059-4463-9942-3c84bd8a556d-kube-api-access-fgkx5\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.706712 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0826aec-f059-4463-9942-3c84bd8a556d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b0826aec-f059-4463-9942-3c84bd8a556d\") " pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.783722 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:05 crc kubenswrapper[4979]: I1204 13:18:05.833749 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.225455 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" path="/var/lib/kubelet/pods/249ad63a-955e-4fe2-b30a-4534dd49f8ef/volumes" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.226401 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5" path="/var/lib/kubelet/pods/665c8ff9-2c3b-42e1-8cc4-dd0f18b5e3c5/volumes" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.266999 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 04 13:18:06 crc kubenswrapper[4979]: W1204 13:18:06.273271 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0826aec_f059_4463_9942_3c84bd8a556d.slice/crio-ebc939ee2c00fe634ffe3a436135482fe798d63bb8a88b462348dab6b3001e6b WatchSource:0}: Error finding container ebc939ee2c00fe634ffe3a436135482fe798d63bb8a88b462348dab6b3001e6b: Status 404 returned error can't find the container with id ebc939ee2c00fe634ffe3a436135482fe798d63bb8a88b462348dab6b3001e6b Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.354292 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b0826aec-f059-4463-9942-3c84bd8a556d","Type":"ContainerStarted","Data":"ebc939ee2c00fe634ffe3a436135482fe798d63bb8a88b462348dab6b3001e6b"} Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.375317 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 13:18:06 crc kubenswrapper[4979]: W1204 13:18:06.375358 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf1145a5_ea36_490c_9360_98feed19a052.slice/crio-24049e61222c2c7adbb5589b7c9c01c329d9b4e387a14f345d2ab440076566c5 WatchSource:0}: Error finding container 24049e61222c2c7adbb5589b7c9c01c329d9b4e387a14f345d2ab440076566c5: Status 404 returned error can't find the container with id 24049e61222c2c7adbb5589b7c9c01c329d9b4e387a14f345d2ab440076566c5 Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.813620 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": read tcp 10.217.0.2:48190->10.217.1.74:8775: read: connection reset by peer" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.813655 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.74:8775/\": read tcp 10.217.0.2:48194->10.217.1.74:8775: read: connection reset by peer" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.869145 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": read tcp 10.217.0.2:51938->10.217.1.73:8774: read: connection reset by peer" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.869509 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.73:8774/\": read tcp 10.217.0.2:51952->10.217.1.73:8774: read: connection reset by peer" Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.933542 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:18:06 crc kubenswrapper[4979]: I1204 13:18:06.933813 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="2d0d52ad-7107-44f7-b317-d3b628ee078e" containerName="nova-cell1-conductor-conductor" containerID="cri-o://e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03" gracePeriod=30 Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.392420 4979 generic.go:334] "Generic (PLEG): container finished" podID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerID="4c9a940c9aefdc19229ce43bd16b871d1dbc397f9bbd271044a98121efd9b57c" exitCode=0 Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.392628 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bda865f5-85ba-4d30-84e7-057bc6e2e1c9","Type":"ContainerDied","Data":"4c9a940c9aefdc19229ce43bd16b871d1dbc397f9bbd271044a98121efd9b57c"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.392918 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bda865f5-85ba-4d30-84e7-057bc6e2e1c9","Type":"ContainerDied","Data":"f575bba795b0e8ea12b20ba1c206db9ebbe7a27ec7509a7011f1708888b8bf79"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.392939 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f575bba795b0e8ea12b20ba1c206db9ebbe7a27ec7509a7011f1708888b8bf79" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.398756 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cf1145a5-ea36-490c-9360-98feed19a052","Type":"ContainerStarted","Data":"642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.398800 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cf1145a5-ea36-490c-9360-98feed19a052","Type":"ContainerStarted","Data":"24049e61222c2c7adbb5589b7c9c01c329d9b4e387a14f345d2ab440076566c5"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.402901 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b0826aec-f059-4463-9942-3c84bd8a556d","Type":"ContainerStarted","Data":"5348b22e99f2e828aa2cc6e99ddef14318b977ad1afd99f48dbb7746126798c5"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.407138 4979 generic.go:334] "Generic (PLEG): container finished" podID="83299115-b631-4e6e-901a-e82944e5c21c" containerID="07bbe03959b56919a54c2cc12b07d0442d8ce804b250d0a6a5d7f7d887d3d15a" exitCode=0 Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.407195 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"83299115-b631-4e6e-901a-e82944e5c21c","Type":"ContainerDied","Data":"07bbe03959b56919a54c2cc12b07d0442d8ce804b250d0a6a5d7f7d887d3d15a"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.410462 4979 generic.go:334] "Generic (PLEG): container finished" podID="e28a7327-643f-45f1-848c-d971f083136a" containerID="8e6430c1953b1184982391fbe8caa201a15a1e05432d4e5e5b07895d2aa2f71d" exitCode=0 Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.410535 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e28a7327-643f-45f1-848c-d971f083136a","Type":"ContainerDied","Data":"8e6430c1953b1184982391fbe8caa201a15a1e05432d4e5e5b07895d2aa2f71d"} Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.424671 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.42464655 podStartE2EDuration="2.42464655s" podCreationTimestamp="2025-12-04 13:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:07.419070552 +0000 UTC m=+5711.693366366" watchObservedRunningTime="2025-12-04 13:18:07.42464655 +0000 UTC m=+5711.698942354" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.440623 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.440603184 podStartE2EDuration="2.440603184s" podCreationTimestamp="2025-12-04 13:18:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:07.437749138 +0000 UTC m=+5711.712044962" watchObservedRunningTime="2025-12-04 13:18:07.440603184 +0000 UTC m=+5711.714898988" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.531665 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.549858 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.628592 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbtpl\" (UniqueName: \"kubernetes.io/projected/e28a7327-643f-45f1-848c-d971f083136a-kube-api-access-gbtpl\") pod \"e28a7327-643f-45f1-848c-d971f083136a\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629230 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-config-data\") pod \"e28a7327-643f-45f1-848c-d971f083136a\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629460 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8l6g\" (UniqueName: \"kubernetes.io/projected/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-kube-api-access-x8l6g\") pod \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629568 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-combined-ca-bundle\") pod \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629678 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-combined-ca-bundle\") pod \"e28a7327-643f-45f1-848c-d971f083136a\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629770 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28a7327-643f-45f1-848c-d971f083136a-logs\") pod \"e28a7327-643f-45f1-848c-d971f083136a\" (UID: \"e28a7327-643f-45f1-848c-d971f083136a\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629869 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-config-data\") pod \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.629987 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-logs\") pod \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\" (UID: \"bda865f5-85ba-4d30-84e7-057bc6e2e1c9\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.633052 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-logs" (OuterVolumeSpecName: "logs") pod "bda865f5-85ba-4d30-84e7-057bc6e2e1c9" (UID: "bda865f5-85ba-4d30-84e7-057bc6e2e1c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.633646 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e28a7327-643f-45f1-848c-d971f083136a-logs" (OuterVolumeSpecName: "logs") pod "e28a7327-643f-45f1-848c-d971f083136a" (UID: "e28a7327-643f-45f1-848c-d971f083136a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.664608 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e28a7327-643f-45f1-848c-d971f083136a-kube-api-access-gbtpl" (OuterVolumeSpecName: "kube-api-access-gbtpl") pod "e28a7327-643f-45f1-848c-d971f083136a" (UID: "e28a7327-643f-45f1-848c-d971f083136a"). InnerVolumeSpecName "kube-api-access-gbtpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.674635 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-kube-api-access-x8l6g" (OuterVolumeSpecName: "kube-api-access-x8l6g") pod "bda865f5-85ba-4d30-84e7-057bc6e2e1c9" (UID: "bda865f5-85ba-4d30-84e7-057bc6e2e1c9"). InnerVolumeSpecName "kube-api-access-x8l6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.701496 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-config-data" (OuterVolumeSpecName: "config-data") pod "e28a7327-643f-45f1-848c-d971f083136a" (UID: "e28a7327-643f-45f1-848c-d971f083136a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.704633 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bda865f5-85ba-4d30-84e7-057bc6e2e1c9" (UID: "bda865f5-85ba-4d30-84e7-057bc6e2e1c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.733261 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8l6g\" (UniqueName: \"kubernetes.io/projected/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-kube-api-access-x8l6g\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.733318 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.733335 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e28a7327-643f-45f1-848c-d971f083136a-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.733350 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.733363 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbtpl\" (UniqueName: \"kubernetes.io/projected/e28a7327-643f-45f1-848c-d971f083136a-kube-api-access-gbtpl\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.733374 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.735496 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-config-data" (OuterVolumeSpecName: "config-data") pod "bda865f5-85ba-4d30-84e7-057bc6e2e1c9" (UID: "bda865f5-85ba-4d30-84e7-057bc6e2e1c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.745842 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.747846 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e28a7327-643f-45f1-848c-d971f083136a" (UID: "e28a7327-643f-45f1-848c-d971f083136a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.835205 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9fxn\" (UniqueName: \"kubernetes.io/projected/83299115-b631-4e6e-901a-e82944e5c21c-kube-api-access-g9fxn\") pod \"83299115-b631-4e6e-901a-e82944e5c21c\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.835289 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-combined-ca-bundle\") pod \"83299115-b631-4e6e-901a-e82944e5c21c\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.835398 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-config-data\") pod \"83299115-b631-4e6e-901a-e82944e5c21c\" (UID: \"83299115-b631-4e6e-901a-e82944e5c21c\") " Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.835977 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bda865f5-85ba-4d30-84e7-057bc6e2e1c9-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.836006 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e28a7327-643f-45f1-848c-d971f083136a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.844853 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83299115-b631-4e6e-901a-e82944e5c21c-kube-api-access-g9fxn" (OuterVolumeSpecName: "kube-api-access-g9fxn") pod "83299115-b631-4e6e-901a-e82944e5c21c" (UID: "83299115-b631-4e6e-901a-e82944e5c21c"). InnerVolumeSpecName "kube-api-access-g9fxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.860626 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83299115-b631-4e6e-901a-e82944e5c21c" (UID: "83299115-b631-4e6e-901a-e82944e5c21c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.863477 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-config-data" (OuterVolumeSpecName: "config-data") pod "83299115-b631-4e6e-901a-e82944e5c21c" (UID: "83299115-b631-4e6e-901a-e82944e5c21c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.938411 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9fxn\" (UniqueName: \"kubernetes.io/projected/83299115-b631-4e6e-901a-e82944e5c21c-kube-api-access-g9fxn\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.938445 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:07 crc kubenswrapper[4979]: I1204 13:18:07.938456 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83299115-b631-4e6e-901a-e82944e5c21c-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.421070 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"83299115-b631-4e6e-901a-e82944e5c21c","Type":"ContainerDied","Data":"e429d201d4cac11ffca9d562a17612ae4cf1c25103f97155aafd910072c1b584"} Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.421081 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.421125 4979 scope.go:117] "RemoveContainer" containerID="07bbe03959b56919a54c2cc12b07d0442d8ce804b250d0a6a5d7f7d887d3d15a" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.424993 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e28a7327-643f-45f1-848c-d971f083136a","Type":"ContainerDied","Data":"e13eb49d8bf633856d3ba3faae306a08cf119b4046907dbf13f8e96ef79c814b"} Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.425146 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.425204 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.449828 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.457132 4979 scope.go:117] "RemoveContainer" containerID="8e6430c1953b1184982391fbe8caa201a15a1e05432d4e5e5b07895d2aa2f71d" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.480876 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.491032 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: E1204 13:18:08.491597 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-log" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.491631 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-log" Dec 04 13:18:08 crc kubenswrapper[4979]: E1204 13:18:08.491651 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-metadata" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.491660 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-metadata" Dec 04 13:18:08 crc kubenswrapper[4979]: E1204 13:18:08.491696 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83299115-b631-4e6e-901a-e82944e5c21c" containerName="nova-cell0-conductor-conductor" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.491705 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="83299115-b631-4e6e-901a-e82944e5c21c" containerName="nova-cell0-conductor-conductor" Dec 04 13:18:08 crc kubenswrapper[4979]: E1204 13:18:08.491724 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-log" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.491733 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-log" Dec 04 13:18:08 crc kubenswrapper[4979]: E1204 13:18:08.491754 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-api" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.491762 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-api" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.492011 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-metadata" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.492033 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-api" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.492064 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="83299115-b631-4e6e-901a-e82944e5c21c" containerName="nova-cell0-conductor-conductor" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.492080 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" containerName="nova-metadata-log" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.492098 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e28a7327-643f-45f1-848c-d971f083136a" containerName="nova-api-log" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.492891 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.496220 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.500163 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.504957 4979 scope.go:117] "RemoveContainer" containerID="6aca61a38fb477a9267cb4b0138b2346e65ac917289490a6259f0aff4c6f5466" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.514485 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.529697 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.536757 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.556359 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.577394 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.579504 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.582893 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.606404 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.608774 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.615682 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.625720 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.645274 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654649 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654691 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654743 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58gsl\" (UniqueName: \"kubernetes.io/projected/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-kube-api-access-58gsl\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654762 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-config-data\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654795 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc6d010c-f547-4ea1-9d88-16a8a85834bf-logs\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654814 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.654849 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgh8j\" (UniqueName: \"kubernetes.io/projected/bc6d010c-f547-4ea1-9d88-16a8a85834bf-kube-api-access-cgh8j\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.755831 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.755906 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.755926 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.755945 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d88d25-879c-461c-a146-89bc738ef3ff-logs\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.755965 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-config-data\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756014 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58gsl\" (UniqueName: \"kubernetes.io/projected/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-kube-api-access-58gsl\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756101 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-config-data\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756215 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc6d010c-f547-4ea1-9d88-16a8a85834bf-logs\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756260 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756403 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgh8j\" (UniqueName: \"kubernetes.io/projected/bc6d010c-f547-4ea1-9d88-16a8a85834bf-kube-api-access-cgh8j\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756475 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqp64\" (UniqueName: \"kubernetes.io/projected/88d88d25-879c-461c-a146-89bc738ef3ff-kube-api-access-rqp64\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.756807 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc6d010c-f547-4ea1-9d88-16a8a85834bf-logs\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.761198 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.761482 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.765996 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.766081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-config-data\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.775892 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58gsl\" (UniqueName: \"kubernetes.io/projected/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-kube-api-access-58gsl\") pod \"nova-cell0-conductor-0\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.786551 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgh8j\" (UniqueName: \"kubernetes.io/projected/bc6d010c-f547-4ea1-9d88-16a8a85834bf-kube-api-access-cgh8j\") pod \"nova-metadata-0\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.825036 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.858408 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqp64\" (UniqueName: \"kubernetes.io/projected/88d88d25-879c-461c-a146-89bc738ef3ff-kube-api-access-rqp64\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.858515 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.858631 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d88d25-879c-461c-a146-89bc738ef3ff-logs\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.858667 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-config-data\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.859357 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d88d25-879c-461c-a146-89bc738ef3ff-logs\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.863692 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.865903 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-config-data\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.877827 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqp64\" (UniqueName: \"kubernetes.io/projected/88d88d25-879c-461c-a146-89bc738ef3ff-kube-api-access-rqp64\") pod \"nova-api-0\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " pod="openstack/nova-api-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.935850 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 13:18:08 crc kubenswrapper[4979]: I1204 13:18:08.954792 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 13:18:09 crc kubenswrapper[4979]: I1204 13:18:09.374767 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 13:18:09 crc kubenswrapper[4979]: I1204 13:18:09.506549 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458","Type":"ContainerStarted","Data":"0b2e18202413d32c534a7833e482c549758aeba1f849bb4345fdfac63af74be5"} Dec 04 13:18:09 crc kubenswrapper[4979]: I1204 13:18:09.527127 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-cell1-novncproxy-0" podUID="249ad63a-955e-4fe2-b30a-4534dd49f8ef" containerName="nova-cell1-novncproxy-novncproxy" probeResult="failure" output="Get \"http://10.217.1.63:6080/vnc_lite.html\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:18:09 crc kubenswrapper[4979]: I1204 13:18:09.571428 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 13:18:09 crc kubenswrapper[4979]: W1204 13:18:09.588913 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88d88d25_879c_461c_a146_89bc738ef3ff.slice/crio-1a14811c51094cd3380d0ab77fb3b5596e1c8f5517083ea416c6ee40c268d4bd WatchSource:0}: Error finding container 1a14811c51094cd3380d0ab77fb3b5596e1c8f5517083ea416c6ee40c268d4bd: Status 404 returned error can't find the container with id 1a14811c51094cd3380d0ab77fb3b5596e1c8f5517083ea416c6ee40c268d4bd Dec 04 13:18:09 crc kubenswrapper[4979]: I1204 13:18:09.601131 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 13:18:09 crc kubenswrapper[4979]: W1204 13:18:09.604978 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc6d010c_f547_4ea1_9d88_16a8a85834bf.slice/crio-fcb5adece712c264be7f9a63d7c7f25cbc9f667c3cdf8e909c277b254831140d WatchSource:0}: Error finding container fcb5adece712c264be7f9a63d7c7f25cbc9f667c3cdf8e909c277b254831140d: Status 404 returned error can't find the container with id fcb5adece712c264be7f9a63d7c7f25cbc9f667c3cdf8e909c277b254831140d Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.199405 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:18:10 crc kubenswrapper[4979]: E1204 13:18:10.199966 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.212868 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83299115-b631-4e6e-901a-e82944e5c21c" path="/var/lib/kubelet/pods/83299115-b631-4e6e-901a-e82944e5c21c/volumes" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.213655 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bda865f5-85ba-4d30-84e7-057bc6e2e1c9" path="/var/lib/kubelet/pods/bda865f5-85ba-4d30-84e7-057bc6e2e1c9/volumes" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.214422 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e28a7327-643f-45f1-848c-d971f083136a" path="/var/lib/kubelet/pods/e28a7327-643f-45f1-848c-d971f083136a/volumes" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.526147 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458","Type":"ContainerStarted","Data":"2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.527594 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.532618 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88d88d25-879c-461c-a146-89bc738ef3ff","Type":"ContainerStarted","Data":"ce04a68a8d5882a2a8474677b476c012dad476be2cac35c4de03abbf3a0a803b"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.532736 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88d88d25-879c-461c-a146-89bc738ef3ff","Type":"ContainerStarted","Data":"e15b275173b8bd8e14ab47413fbf97633b85db28701be0e8cbdb009087d43c9f"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.532756 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88d88d25-879c-461c-a146-89bc738ef3ff","Type":"ContainerStarted","Data":"1a14811c51094cd3380d0ab77fb3b5596e1c8f5517083ea416c6ee40c268d4bd"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.535763 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc6d010c-f547-4ea1-9d88-16a8a85834bf","Type":"ContainerStarted","Data":"d0a361d3f246a9a0b46fcb15732158eba4b629166f2aa067d3ab69b289653692"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.535808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc6d010c-f547-4ea1-9d88-16a8a85834bf","Type":"ContainerStarted","Data":"9f1eda5655293e5c52af5fb24746e23eec2651f8e849d13062d637d2cf6d7f81"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.535819 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc6d010c-f547-4ea1-9d88-16a8a85834bf","Type":"ContainerStarted","Data":"fcb5adece712c264be7f9a63d7c7f25cbc9f667c3cdf8e909c277b254831140d"} Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.561484 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.561462918 podStartE2EDuration="2.561462918s" podCreationTimestamp="2025-12-04 13:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:10.551902585 +0000 UTC m=+5714.826198389" watchObservedRunningTime="2025-12-04 13:18:10.561462918 +0000 UTC m=+5714.835758722" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.596513 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.596496469 podStartE2EDuration="2.596496469s" podCreationTimestamp="2025-12-04 13:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:10.59048801 +0000 UTC m=+5714.864783824" watchObservedRunningTime="2025-12-04 13:18:10.596496469 +0000 UTC m=+5714.870792273" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.597043 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.597036874 podStartE2EDuration="2.597036874s" podCreationTimestamp="2025-12-04 13:18:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:10.572098951 +0000 UTC m=+5714.846394755" watchObservedRunningTime="2025-12-04 13:18:10.597036874 +0000 UTC m=+5714.871332678" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.791668 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:10 crc kubenswrapper[4979]: I1204 13:18:10.834594 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.110003 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.204057 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-combined-ca-bundle\") pod \"2d0d52ad-7107-44f7-b317-d3b628ee078e\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.204255 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hk6rv\" (UniqueName: \"kubernetes.io/projected/2d0d52ad-7107-44f7-b317-d3b628ee078e-kube-api-access-hk6rv\") pod \"2d0d52ad-7107-44f7-b317-d3b628ee078e\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.204348 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-config-data\") pod \"2d0d52ad-7107-44f7-b317-d3b628ee078e\" (UID: \"2d0d52ad-7107-44f7-b317-d3b628ee078e\") " Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.211222 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d0d52ad-7107-44f7-b317-d3b628ee078e-kube-api-access-hk6rv" (OuterVolumeSpecName: "kube-api-access-hk6rv") pod "2d0d52ad-7107-44f7-b317-d3b628ee078e" (UID: "2d0d52ad-7107-44f7-b317-d3b628ee078e"). InnerVolumeSpecName "kube-api-access-hk6rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.231946 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d0d52ad-7107-44f7-b317-d3b628ee078e" (UID: "2d0d52ad-7107-44f7-b317-d3b628ee078e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.241447 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-config-data" (OuterVolumeSpecName: "config-data") pod "2d0d52ad-7107-44f7-b317-d3b628ee078e" (UID: "2d0d52ad-7107-44f7-b317-d3b628ee078e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.309658 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.309686 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d0d52ad-7107-44f7-b317-d3b628ee078e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.309696 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hk6rv\" (UniqueName: \"kubernetes.io/projected/2d0d52ad-7107-44f7-b317-d3b628ee078e-kube-api-access-hk6rv\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.547854 4979 generic.go:334] "Generic (PLEG): container finished" podID="2d0d52ad-7107-44f7-b317-d3b628ee078e" containerID="e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03" exitCode=0 Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.547901 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.547918 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0d52ad-7107-44f7-b317-d3b628ee078e","Type":"ContainerDied","Data":"e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03"} Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.548321 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"2d0d52ad-7107-44f7-b317-d3b628ee078e","Type":"ContainerDied","Data":"f3314bc7409f9cf1aaa693041c84d9f872ce68f7c3a376d966e3500eb87d6327"} Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.548345 4979 scope.go:117] "RemoveContainer" containerID="e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.572272 4979 scope.go:117] "RemoveContainer" containerID="e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03" Dec 04 13:18:11 crc kubenswrapper[4979]: E1204 13:18:11.572734 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03\": container with ID starting with e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03 not found: ID does not exist" containerID="e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.572779 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03"} err="failed to get container status \"e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03\": rpc error: code = NotFound desc = could not find container \"e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03\": container with ID starting with e6433e549f2fc90be01ede2d0d00334e05cf61a5606d904dabf5a59da1183d03 not found: ID does not exist" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.586539 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.595345 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.610995 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:18:11 crc kubenswrapper[4979]: E1204 13:18:11.611546 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d0d52ad-7107-44f7-b317-d3b628ee078e" containerName="nova-cell1-conductor-conductor" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.611571 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d0d52ad-7107-44f7-b317-d3b628ee078e" containerName="nova-cell1-conductor-conductor" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.611800 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d0d52ad-7107-44f7-b317-d3b628ee078e" containerName="nova-cell1-conductor-conductor" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.612642 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.617724 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.626699 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.716135 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.716230 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.716318 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgm2r\" (UniqueName: \"kubernetes.io/projected/235f670f-55af-4667-998a-9d0066bc01d1-kube-api-access-lgm2r\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.818127 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.818217 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.818279 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgm2r\" (UniqueName: \"kubernetes.io/projected/235f670f-55af-4667-998a-9d0066bc01d1-kube-api-access-lgm2r\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.823930 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.830282 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.835218 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgm2r\" (UniqueName: \"kubernetes.io/projected/235f670f-55af-4667-998a-9d0066bc01d1-kube-api-access-lgm2r\") pod \"nova-cell1-conductor-0\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:11 crc kubenswrapper[4979]: I1204 13:18:11.932448 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:12 crc kubenswrapper[4979]: I1204 13:18:12.211637 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d0d52ad-7107-44f7-b317-d3b628ee078e" path="/var/lib/kubelet/pods/2d0d52ad-7107-44f7-b317-d3b628ee078e/volumes" Dec 04 13:18:12 crc kubenswrapper[4979]: I1204 13:18:12.371070 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 13:18:12 crc kubenswrapper[4979]: I1204 13:18:12.559033 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"235f670f-55af-4667-998a-9d0066bc01d1","Type":"ContainerStarted","Data":"333398ff21ea6229bcc352dee8bd6343d092c0921c76c3e6c5cdf1e68245ff56"} Dec 04 13:18:13 crc kubenswrapper[4979]: I1204 13:18:13.569424 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"235f670f-55af-4667-998a-9d0066bc01d1","Type":"ContainerStarted","Data":"1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8"} Dec 04 13:18:13 crc kubenswrapper[4979]: I1204 13:18:13.570503 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:13 crc kubenswrapper[4979]: I1204 13:18:13.590092 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.590071639 podStartE2EDuration="2.590071639s" podCreationTimestamp="2025-12-04 13:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:13.58519424 +0000 UTC m=+5717.859490044" watchObservedRunningTime="2025-12-04 13:18:13.590071639 +0000 UTC m=+5717.864367443" Dec 04 13:18:13 crc kubenswrapper[4979]: I1204 13:18:13.937013 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:18:13 crc kubenswrapper[4979]: I1204 13:18:13.937353 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 13:18:15 crc kubenswrapper[4979]: I1204 13:18:15.784932 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:15 crc kubenswrapper[4979]: I1204 13:18:15.798591 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:15 crc kubenswrapper[4979]: I1204 13:18:15.834691 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 13:18:15 crc kubenswrapper[4979]: I1204 13:18:15.867892 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 13:18:16 crc kubenswrapper[4979]: I1204 13:18:16.605833 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 04 13:18:16 crc kubenswrapper[4979]: I1204 13:18:16.630958 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 13:18:18 crc kubenswrapper[4979]: I1204 13:18:18.850173 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 13:18:18 crc kubenswrapper[4979]: I1204 13:18:18.936991 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 13:18:18 crc kubenswrapper[4979]: I1204 13:18:18.937085 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 13:18:18 crc kubenswrapper[4979]: I1204 13:18:18.955474 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 13:18:18 crc kubenswrapper[4979]: I1204 13:18:18.955531 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 13:18:20 crc kubenswrapper[4979]: I1204 13:18:20.101773 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.86:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:18:20 crc kubenswrapper[4979]: I1204 13:18:20.101783 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.86:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:18:20 crc kubenswrapper[4979]: I1204 13:18:20.101812 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:18:20 crc kubenswrapper[4979]: I1204 13:18:20.101777 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 13:18:21 crc kubenswrapper[4979]: I1204 13:18:21.961055 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.201341 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:18:23 crc kubenswrapper[4979]: E1204 13:18:23.201762 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.208112 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.211482 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.213925 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.221755 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.231868 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.231917 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.231968 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b36c1245-6472-4a5a-8739-a65af0f3ecea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.232006 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hjv\" (UniqueName: \"kubernetes.io/projected/b36c1245-6472-4a5a-8739-a65af0f3ecea-kube-api-access-87hjv\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.232165 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-scripts\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.232210 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.332936 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-scripts\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.332990 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.333038 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.333053 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.333081 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b36c1245-6472-4a5a-8739-a65af0f3ecea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.333104 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hjv\" (UniqueName: \"kubernetes.io/projected/b36c1245-6472-4a5a-8739-a65af0f3ecea-kube-api-access-87hjv\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.333445 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b36c1245-6472-4a5a-8739-a65af0f3ecea-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.339697 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-scripts\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.339997 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.340136 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.342274 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.352806 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hjv\" (UniqueName: \"kubernetes.io/projected/b36c1245-6472-4a5a-8739-a65af0f3ecea-kube-api-access-87hjv\") pod \"cinder-scheduler-0\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:23 crc kubenswrapper[4979]: I1204 13:18:23.559357 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 13:18:24 crc kubenswrapper[4979]: I1204 13:18:24.008119 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:24 crc kubenswrapper[4979]: W1204 13:18:24.018042 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb36c1245_6472_4a5a_8739_a65af0f3ecea.slice/crio-9eb57d09641a2d75a086ec799d9fad5460f400aeeae002fcd7f7cf62c81b64bb WatchSource:0}: Error finding container 9eb57d09641a2d75a086ec799d9fad5460f400aeeae002fcd7f7cf62c81b64bb: Status 404 returned error can't find the container with id 9eb57d09641a2d75a086ec799d9fad5460f400aeeae002fcd7f7cf62c81b64bb Dec 04 13:18:24 crc kubenswrapper[4979]: I1204 13:18:24.621557 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:18:24 crc kubenswrapper[4979]: I1204 13:18:24.622109 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api-log" containerID="cri-o://f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1" gracePeriod=30 Dec 04 13:18:24 crc kubenswrapper[4979]: I1204 13:18:24.622361 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api" containerID="cri-o://ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1" gracePeriod=30 Dec 04 13:18:24 crc kubenswrapper[4979]: I1204 13:18:24.680233 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b36c1245-6472-4a5a-8739-a65af0f3ecea","Type":"ContainerStarted","Data":"9eb57d09641a2d75a086ec799d9fad5460f400aeeae002fcd7f7cf62c81b64bb"} Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.166329 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.168461 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.171815 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.191383 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291622 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291693 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291721 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291744 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291776 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-run\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291798 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291841 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291875 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291913 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291965 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.291996 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.292077 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.292104 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.292159 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.292184 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjg28\" (UniqueName: \"kubernetes.io/projected/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-kube-api-access-bjg28\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.292230 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.394352 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.394423 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.394464 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.394560 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.394739 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395162 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjg28\" (UniqueName: \"kubernetes.io/projected/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-kube-api-access-bjg28\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395349 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395400 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395444 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395480 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395503 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395551 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-run\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395578 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395604 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395617 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395676 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395722 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395768 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395809 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.395885 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-dev\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.396035 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-run\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.396078 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.396157 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.396161 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-sys\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.396272 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.396430 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.402050 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.402091 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.402382 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.402816 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.408355 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.421065 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjg28\" (UniqueName: \"kubernetes.io/projected/a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a-kube-api-access-bjg28\") pod \"cinder-volume-volume1-0\" (UID: \"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a\") " pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.508709 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.713803 4979 generic.go:334] "Generic (PLEG): container finished" podID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerID="f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1" exitCode=143 Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.713956 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbdd03c8-bb67-4284-bc45-be2401c5d665","Type":"ContainerDied","Data":"f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1"} Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.717398 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b36c1245-6472-4a5a-8739-a65af0f3ecea","Type":"ContainerStarted","Data":"7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134"} Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.717466 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b36c1245-6472-4a5a-8739-a65af0f3ecea","Type":"ContainerStarted","Data":"27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4"} Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.755446 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=2.755419009 podStartE2EDuration="2.755419009s" podCreationTimestamp="2025-12-04 13:18:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:25.748526036 +0000 UTC m=+5730.022821850" watchObservedRunningTime="2025-12-04 13:18:25.755419009 +0000 UTC m=+5730.029714823" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.842981 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.846459 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.849292 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.856990 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915023 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915100 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-dev\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915127 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c98898f0-b483-4f47-9b32-101da166e4b7-ceph\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915159 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-run\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915223 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915269 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915315 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-lib-modules\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915350 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915381 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-sys\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915409 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-config-data\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915437 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915465 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915505 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7flpq\" (UniqueName: \"kubernetes.io/projected/c98898f0-b483-4f47-9b32-101da166e4b7-kube-api-access-7flpq\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915556 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-scripts\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:25 crc kubenswrapper[4979]: I1204 13:18:25.915576 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.017472 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-sys\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.017291 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-sys\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.017999 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-config-data\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018035 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018072 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018117 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7flpq\" (UniqueName: \"kubernetes.io/projected/c98898f0-b483-4f47-9b32-101da166e4b7-kube-api-access-7flpq\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018145 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018164 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-scripts\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018184 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018258 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018326 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-dev\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018354 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c98898f0-b483-4f47-9b32-101da166e4b7-ceph\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018386 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-run\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018419 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018467 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018497 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-lib-modules\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018536 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.018654 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.019638 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-nvme\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.019728 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.019769 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.020106 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.020760 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-run\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.020831 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-dev\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.025629 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/c98898f0-b483-4f47-9b32-101da166e4b7-ceph\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.025868 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.025945 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/c98898f0-b483-4f47-9b32-101da166e4b7-lib-modules\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.028231 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-config-data\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.028749 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-scripts\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.032010 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.033148 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c98898f0-b483-4f47-9b32-101da166e4b7-config-data-custom\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.043940 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7flpq\" (UniqueName: \"kubernetes.io/projected/c98898f0-b483-4f47-9b32-101da166e4b7-kube-api-access-7flpq\") pod \"cinder-backup-0\" (UID: \"c98898f0-b483-4f47-9b32-101da166e4b7\") " pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: W1204 13:18:26.126734 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda76a87b5_d6c7_4b1a_92b1_6aad4febdc9a.slice/crio-2d7cf66a1e5da5dea572bfa4817c0c2d6e2fcceac765f404169c1162bfd06370 WatchSource:0}: Error finding container 2d7cf66a1e5da5dea572bfa4817c0c2d6e2fcceac765f404169c1162bfd06370: Status 404 returned error can't find the container with id 2d7cf66a1e5da5dea572bfa4817c0c2d6e2fcceac765f404169c1162bfd06370 Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.137987 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.178376 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.735978 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a","Type":"ContainerStarted","Data":"2d7cf66a1e5da5dea572bfa4817c0c2d6e2fcceac765f404169c1162bfd06370"} Dec 04 13:18:26 crc kubenswrapper[4979]: I1204 13:18:26.779969 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 04 13:18:27 crc kubenswrapper[4979]: I1204 13:18:27.754921 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a","Type":"ContainerStarted","Data":"e537b6660eb33f6a3cf3f7b69157934d61e5663d41bc8adb56b8ab9e8ecef13f"} Dec 04 13:18:27 crc kubenswrapper[4979]: I1204 13:18:27.755481 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a","Type":"ContainerStarted","Data":"29adf701af1dec3d64d756f5bcba427959ff4572ace3d778bcd4ec7d376fb1db"} Dec 04 13:18:27 crc kubenswrapper[4979]: I1204 13:18:27.759761 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c98898f0-b483-4f47-9b32-101da166e4b7","Type":"ContainerStarted","Data":"a8087e71a9d1d8511458c64f897c96db2719190d000daef2ffa10ecdf3b23537"} Dec 04 13:18:27 crc kubenswrapper[4979]: I1204 13:18:27.772207 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.1.82:8776/healthcheck\": read tcp 10.217.0.2:53848->10.217.1.82:8776: read: connection reset by peer" Dec 04 13:18:27 crc kubenswrapper[4979]: I1204 13:18:27.790960 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=1.991225682 podStartE2EDuration="2.790931421s" podCreationTimestamp="2025-12-04 13:18:25 +0000 UTC" firstStartedPulling="2025-12-04 13:18:26.128853307 +0000 UTC m=+5730.403149111" lastFinishedPulling="2025-12-04 13:18:26.928559046 +0000 UTC m=+5731.202854850" observedRunningTime="2025-12-04 13:18:27.783566075 +0000 UTC m=+5732.057861879" watchObservedRunningTime="2025-12-04 13:18:27.790931421 +0000 UTC m=+5732.065227245" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.210875 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.278642 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-scripts\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.278771 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data-custom\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.278896 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbdd03c8-bb67-4284-bc45-be2401c5d665-logs\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.278933 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbdd03c8-bb67-4284-bc45-be2401c5d665-etc-machine-id\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.278956 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-combined-ca-bundle\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.279057 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.279201 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wt9j9\" (UniqueName: \"kubernetes.io/projected/dbdd03c8-bb67-4284-bc45-be2401c5d665-kube-api-access-wt9j9\") pod \"dbdd03c8-bb67-4284-bc45-be2401c5d665\" (UID: \"dbdd03c8-bb67-4284-bc45-be2401c5d665\") " Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.280136 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dbdd03c8-bb67-4284-bc45-be2401c5d665-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.280525 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbdd03c8-bb67-4284-bc45-be2401c5d665-logs" (OuterVolumeSpecName: "logs") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.283788 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbdd03c8-bb67-4284-bc45-be2401c5d665-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.283826 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/dbdd03c8-bb67-4284-bc45-be2401c5d665-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.287254 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbdd03c8-bb67-4284-bc45-be2401c5d665-kube-api-access-wt9j9" (OuterVolumeSpecName: "kube-api-access-wt9j9") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "kube-api-access-wt9j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.290679 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-scripts" (OuterVolumeSpecName: "scripts") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.295191 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.315499 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.362913 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data" (OuterVolumeSpecName: "config-data") pod "dbdd03c8-bb67-4284-bc45-be2401c5d665" (UID: "dbdd03c8-bb67-4284-bc45-be2401c5d665"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.385267 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wt9j9\" (UniqueName: \"kubernetes.io/projected/dbdd03c8-bb67-4284-bc45-be2401c5d665-kube-api-access-wt9j9\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.385329 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.385339 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.385348 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.385357 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbdd03c8-bb67-4284-bc45-be2401c5d665-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.560037 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.776608 4979 generic.go:334] "Generic (PLEG): container finished" podID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerID="ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1" exitCode=0 Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.776858 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbdd03c8-bb67-4284-bc45-be2401c5d665","Type":"ContainerDied","Data":"ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1"} Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.776892 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"dbdd03c8-bb67-4284-bc45-be2401c5d665","Type":"ContainerDied","Data":"ef5de585b1c540676a20bbc2cc2f5cdbc6d49ffb9b1a13facf2f5a638e5d4eff"} Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.776835 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.776913 4979 scope.go:117] "RemoveContainer" containerID="ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.783130 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c98898f0-b483-4f47-9b32-101da166e4b7","Type":"ContainerStarted","Data":"b92103202426e668aae675834217746923b5cffe5a48e81083b2eb513f63b88a"} Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.783170 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"c98898f0-b483-4f47-9b32-101da166e4b7","Type":"ContainerStarted","Data":"ae95f54f00dbcc44596f353668f34dc814c45187896e62d9ce70aac0a0ee88a6"} Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.810374 4979 scope.go:117] "RemoveContainer" containerID="f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.813753 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.019407394 podStartE2EDuration="3.81373294s" podCreationTimestamp="2025-12-04 13:18:25 +0000 UTC" firstStartedPulling="2025-12-04 13:18:26.86854782 +0000 UTC m=+5731.142843614" lastFinishedPulling="2025-12-04 13:18:27.662873356 +0000 UTC m=+5731.937169160" observedRunningTime="2025-12-04 13:18:28.811010928 +0000 UTC m=+5733.085306742" watchObservedRunningTime="2025-12-04 13:18:28.81373294 +0000 UTC m=+5733.088028764" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.857710 4979 scope.go:117] "RemoveContainer" containerID="ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1" Dec 04 13:18:28 crc kubenswrapper[4979]: E1204 13:18:28.858425 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1\": container with ID starting with ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1 not found: ID does not exist" containerID="ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.858601 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1"} err="failed to get container status \"ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1\": rpc error: code = NotFound desc = could not find container \"ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1\": container with ID starting with ecddc11658318c1d4fee9681f66a256331f32a0429b317528ed0eb1aa4cb4bf1 not found: ID does not exist" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.858720 4979 scope.go:117] "RemoveContainer" containerID="f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1" Dec 04 13:18:28 crc kubenswrapper[4979]: E1204 13:18:28.861042 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1\": container with ID starting with f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1 not found: ID does not exist" containerID="f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.861084 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1"} err="failed to get container status \"f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1\": rpc error: code = NotFound desc = could not find container \"f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1\": container with ID starting with f82dad567fe3a00157545ec7516db2d2d551ca36b7d20c8c400fa69da7ebf0b1 not found: ID does not exist" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.876291 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.890000 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.901284 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:18:28 crc kubenswrapper[4979]: E1204 13:18:28.902210 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.902263 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api" Dec 04 13:18:28 crc kubenswrapper[4979]: E1204 13:18:28.902286 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api-log" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.902384 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api-log" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.902844 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.902861 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" containerName="cinder-api-log" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.905686 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.908783 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.922989 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.940889 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.941568 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.944643 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.966686 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.967887 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.969399 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.969450 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.981677 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.983014 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.995775 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.996040 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-config-data-custom\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.996216 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-logs\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.996360 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.996509 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-config-data\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.996688 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-scripts\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:28 crc kubenswrapper[4979]: I1204 13:18:28.996802 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z2q6\" (UniqueName: \"kubernetes.io/projected/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-kube-api-access-7z2q6\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098405 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-config-data-custom\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098494 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-logs\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098523 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098560 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-config-data\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098588 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-etc-machine-id\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098649 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-scripts\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.098680 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z2q6\" (UniqueName: \"kubernetes.io/projected/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-kube-api-access-7z2q6\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.099072 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-logs\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.106607 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.106643 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-scripts\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.112549 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-config-data\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.115885 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-config-data-custom\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.116744 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z2q6\" (UniqueName: \"kubernetes.io/projected/96fc371f-833c-45a0-81c6-b91bfb6cb1c0-kube-api-access-7z2q6\") pod \"cinder-api-0\" (UID: \"96fc371f-833c-45a0-81c6-b91bfb6cb1c0\") " pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.238663 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.725958 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.807076 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96fc371f-833c-45a0-81c6-b91bfb6cb1c0","Type":"ContainerStarted","Data":"dcf31d83bd3a8c1b1b86472d84700a324f60e7b8d1a80fe321608066f5de6110"} Dec 04 13:18:29 crc kubenswrapper[4979]: I1204 13:18:29.811194 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 13:18:30 crc kubenswrapper[4979]: I1204 13:18:30.211742 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbdd03c8-bb67-4284-bc45-be2401c5d665" path="/var/lib/kubelet/pods/dbdd03c8-bb67-4284-bc45-be2401c5d665/volumes" Dec 04 13:18:30 crc kubenswrapper[4979]: I1204 13:18:30.509353 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:30 crc kubenswrapper[4979]: I1204 13:18:30.826252 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96fc371f-833c-45a0-81c6-b91bfb6cb1c0","Type":"ContainerStarted","Data":"00fc1c200667bb5efb03411bc631c14f2e1e8567c9b8c8aff738cc3564d47da6"} Dec 04 13:18:31 crc kubenswrapper[4979]: I1204 13:18:31.179844 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 04 13:18:31 crc kubenswrapper[4979]: I1204 13:18:31.842575 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"96fc371f-833c-45a0-81c6-b91bfb6cb1c0","Type":"ContainerStarted","Data":"843fc8375a5c7d497895f0f15fd31c50ceba3a1cf3290cd0c58b0bb56f380e95"} Dec 04 13:18:31 crc kubenswrapper[4979]: I1204 13:18:31.842923 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 04 13:18:31 crc kubenswrapper[4979]: I1204 13:18:31.866061 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.866037082 podStartE2EDuration="3.866037082s" podCreationTimestamp="2025-12-04 13:18:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:31.859331544 +0000 UTC m=+5736.133627358" watchObservedRunningTime="2025-12-04 13:18:31.866037082 +0000 UTC m=+5736.140332886" Dec 04 13:18:33 crc kubenswrapper[4979]: I1204 13:18:33.792147 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 13:18:33 crc kubenswrapper[4979]: I1204 13:18:33.852992 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:33 crc kubenswrapper[4979]: I1204 13:18:33.860045 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="cinder-scheduler" containerID="cri-o://27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4" gracePeriod=30 Dec 04 13:18:33 crc kubenswrapper[4979]: I1204 13:18:33.860287 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="probe" containerID="cri-o://7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134" gracePeriod=30 Dec 04 13:18:34 crc kubenswrapper[4979]: I1204 13:18:34.869716 4979 generic.go:334] "Generic (PLEG): container finished" podID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerID="7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134" exitCode=0 Dec 04 13:18:34 crc kubenswrapper[4979]: I1204 13:18:34.869760 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b36c1245-6472-4a5a-8739-a65af0f3ecea","Type":"ContainerDied","Data":"7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134"} Dec 04 13:18:35 crc kubenswrapper[4979]: I1204 13:18:35.726488 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.208956 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:18:36 crc kubenswrapper[4979]: E1204 13:18:36.209844 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.391431 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.596564 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643233 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b36c1245-6472-4a5a-8739-a65af0f3ecea-etc-machine-id\") pod \"b36c1245-6472-4a5a-8739-a65af0f3ecea\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643427 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87hjv\" (UniqueName: \"kubernetes.io/projected/b36c1245-6472-4a5a-8739-a65af0f3ecea-kube-api-access-87hjv\") pod \"b36c1245-6472-4a5a-8739-a65af0f3ecea\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643465 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-scripts\") pod \"b36c1245-6472-4a5a-8739-a65af0f3ecea\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643488 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b36c1245-6472-4a5a-8739-a65af0f3ecea-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b36c1245-6472-4a5a-8739-a65af0f3ecea" (UID: "b36c1245-6472-4a5a-8739-a65af0f3ecea"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643519 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-combined-ca-bundle\") pod \"b36c1245-6472-4a5a-8739-a65af0f3ecea\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643591 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data-custom\") pod \"b36c1245-6472-4a5a-8739-a65af0f3ecea\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.643639 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data\") pod \"b36c1245-6472-4a5a-8739-a65af0f3ecea\" (UID: \"b36c1245-6472-4a5a-8739-a65af0f3ecea\") " Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.644180 4979 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b36c1245-6472-4a5a-8739-a65af0f3ecea-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.651101 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b36c1245-6472-4a5a-8739-a65af0f3ecea" (UID: "b36c1245-6472-4a5a-8739-a65af0f3ecea"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.651740 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b36c1245-6472-4a5a-8739-a65af0f3ecea-kube-api-access-87hjv" (OuterVolumeSpecName: "kube-api-access-87hjv") pod "b36c1245-6472-4a5a-8739-a65af0f3ecea" (UID: "b36c1245-6472-4a5a-8739-a65af0f3ecea"). InnerVolumeSpecName "kube-api-access-87hjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.653058 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-scripts" (OuterVolumeSpecName: "scripts") pod "b36c1245-6472-4a5a-8739-a65af0f3ecea" (UID: "b36c1245-6472-4a5a-8739-a65af0f3ecea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.720260 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b36c1245-6472-4a5a-8739-a65af0f3ecea" (UID: "b36c1245-6472-4a5a-8739-a65af0f3ecea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.746647 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87hjv\" (UniqueName: \"kubernetes.io/projected/b36c1245-6472-4a5a-8739-a65af0f3ecea-kube-api-access-87hjv\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.746692 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.746704 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.746715 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.756461 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data" (OuterVolumeSpecName: "config-data") pod "b36c1245-6472-4a5a-8739-a65af0f3ecea" (UID: "b36c1245-6472-4a5a-8739-a65af0f3ecea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.850372 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b36c1245-6472-4a5a-8739-a65af0f3ecea-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.917445 4979 generic.go:334] "Generic (PLEG): container finished" podID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerID="27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4" exitCode=0 Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.917559 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.917901 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b36c1245-6472-4a5a-8739-a65af0f3ecea","Type":"ContainerDied","Data":"27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4"} Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.918031 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b36c1245-6472-4a5a-8739-a65af0f3ecea","Type":"ContainerDied","Data":"9eb57d09641a2d75a086ec799d9fad5460f400aeeae002fcd7f7cf62c81b64bb"} Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.918108 4979 scope.go:117] "RemoveContainer" containerID="7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.943209 4979 scope.go:117] "RemoveContainer" containerID="27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.954248 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.963846 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.980539 4979 scope.go:117] "RemoveContainer" containerID="7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134" Dec 04 13:18:36 crc kubenswrapper[4979]: E1204 13:18:36.981077 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134\": container with ID starting with 7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134 not found: ID does not exist" containerID="7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.981120 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134"} err="failed to get container status \"7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134\": rpc error: code = NotFound desc = could not find container \"7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134\": container with ID starting with 7b304ea400750e5621fe89da19b2c70690499c9cd9318385a1caaa5e48a45134 not found: ID does not exist" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.981152 4979 scope.go:117] "RemoveContainer" containerID="27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4" Dec 04 13:18:36 crc kubenswrapper[4979]: E1204 13:18:36.981571 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4\": container with ID starting with 27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4 not found: ID does not exist" containerID="27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.981603 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4"} err="failed to get container status \"27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4\": rpc error: code = NotFound desc = could not find container \"27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4\": container with ID starting with 27beba513532a569b2471bb46bcb5ca4a1e9ecfdc014ca9cc713da48ed0b41d4 not found: ID does not exist" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.989032 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:36 crc kubenswrapper[4979]: E1204 13:18:36.989698 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="probe" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.989783 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="probe" Dec 04 13:18:36 crc kubenswrapper[4979]: E1204 13:18:36.989883 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="cinder-scheduler" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.989948 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="cinder-scheduler" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.990257 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="probe" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.990382 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" containerName="cinder-scheduler" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.991534 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 13:18:36 crc kubenswrapper[4979]: I1204 13:18:36.993923 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.005072 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.054070 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cndzq\" (UniqueName: \"kubernetes.io/projected/3609683f-00ae-4919-8220-38c41870c261-kube-api-access-cndzq\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.054113 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.054147 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3609683f-00ae-4919-8220-38c41870c261-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.054366 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.054432 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-scripts\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.054483 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-config-data\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.156996 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.157101 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-scripts\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.157150 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-config-data\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.157243 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cndzq\" (UniqueName: \"kubernetes.io/projected/3609683f-00ae-4919-8220-38c41870c261-kube-api-access-cndzq\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.157270 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.157316 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3609683f-00ae-4919-8220-38c41870c261-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.157441 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3609683f-00ae-4919-8220-38c41870c261-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.161251 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-scripts\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.161557 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.161832 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-config-data\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.161841 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3609683f-00ae-4919-8220-38c41870c261-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.177941 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cndzq\" (UniqueName: \"kubernetes.io/projected/3609683f-00ae-4919-8220-38c41870c261-kube-api-access-cndzq\") pod \"cinder-scheduler-0\" (UID: \"3609683f-00ae-4919-8220-38c41870c261\") " pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.322639 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.774113 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 04 13:18:37 crc kubenswrapper[4979]: W1204 13:18:37.774907 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3609683f_00ae_4919_8220_38c41870c261.slice/crio-0d18056938164756a128c3700ca19483fb6d313bc076d613a7260547b3351cd2 WatchSource:0}: Error finding container 0d18056938164756a128c3700ca19483fb6d313bc076d613a7260547b3351cd2: Status 404 returned error can't find the container with id 0d18056938164756a128c3700ca19483fb6d313bc076d613a7260547b3351cd2 Dec 04 13:18:37 crc kubenswrapper[4979]: I1204 13:18:37.929571 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3609683f-00ae-4919-8220-38c41870c261","Type":"ContainerStarted","Data":"0d18056938164756a128c3700ca19483fb6d313bc076d613a7260547b3351cd2"} Dec 04 13:18:38 crc kubenswrapper[4979]: I1204 13:18:38.211971 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b36c1245-6472-4a5a-8739-a65af0f3ecea" path="/var/lib/kubelet/pods/b36c1245-6472-4a5a-8739-a65af0f3ecea/volumes" Dec 04 13:18:38 crc kubenswrapper[4979]: I1204 13:18:38.963648 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3609683f-00ae-4919-8220-38c41870c261","Type":"ContainerStarted","Data":"63e33677f1218140243072d1ee808e28cda9c7ca181db3c34aba901c1631824c"} Dec 04 13:18:39 crc kubenswrapper[4979]: I1204 13:18:39.973013 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3609683f-00ae-4919-8220-38c41870c261","Type":"ContainerStarted","Data":"a0f9c910dccd393ecca32b6bb2969bdfc5816f05b907fa72ea672bc8a5e888b4"} Dec 04 13:18:41 crc kubenswrapper[4979]: I1204 13:18:41.224322 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 04 13:18:41 crc kubenswrapper[4979]: I1204 13:18:41.241678 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.241658251 podStartE2EDuration="5.241658251s" podCreationTimestamp="2025-12-04 13:18:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:18:40.00178635 +0000 UTC m=+5744.276082154" watchObservedRunningTime="2025-12-04 13:18:41.241658251 +0000 UTC m=+5745.515954055" Dec 04 13:18:42 crc kubenswrapper[4979]: I1204 13:18:42.328378 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 04 13:18:47 crc kubenswrapper[4979]: I1204 13:18:47.525913 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 04 13:18:51 crc kubenswrapper[4979]: I1204 13:18:51.198744 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:18:51 crc kubenswrapper[4979]: E1204 13:18:51.199259 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.570171 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8csmw"] Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.572545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.589493 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8csmw"] Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.672642 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-catalog-content\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.672737 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kprn8\" (UniqueName: \"kubernetes.io/projected/75a175f5-76d5-4a9e-920b-2b092642b9d1-kube-api-access-kprn8\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.673018 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-utilities\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.774854 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-utilities\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.774925 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-catalog-content\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.774994 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kprn8\" (UniqueName: \"kubernetes.io/projected/75a175f5-76d5-4a9e-920b-2b092642b9d1-kube-api-access-kprn8\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.775403 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-utilities\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.775422 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-catalog-content\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.796111 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kprn8\" (UniqueName: \"kubernetes.io/projected/75a175f5-76d5-4a9e-920b-2b092642b9d1-kube-api-access-kprn8\") pod \"community-operators-8csmw\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:56 crc kubenswrapper[4979]: I1204 13:18:56.906802 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:18:57 crc kubenswrapper[4979]: I1204 13:18:57.434152 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8csmw"] Dec 04 13:18:57 crc kubenswrapper[4979]: W1204 13:18:57.437730 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75a175f5_76d5_4a9e_920b_2b092642b9d1.slice/crio-333408ad9464c226acc3ca7c004d7e02d1d17edd73f2b9323b54c7c0a937b5fe WatchSource:0}: Error finding container 333408ad9464c226acc3ca7c004d7e02d1d17edd73f2b9323b54c7c0a937b5fe: Status 404 returned error can't find the container with id 333408ad9464c226acc3ca7c004d7e02d1d17edd73f2b9323b54c7c0a937b5fe Dec 04 13:18:58 crc kubenswrapper[4979]: I1204 13:18:58.151652 4979 generic.go:334] "Generic (PLEG): container finished" podID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerID="8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098" exitCode=0 Dec 04 13:18:58 crc kubenswrapper[4979]: I1204 13:18:58.151723 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8csmw" event={"ID":"75a175f5-76d5-4a9e-920b-2b092642b9d1","Type":"ContainerDied","Data":"8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098"} Dec 04 13:18:58 crc kubenswrapper[4979]: I1204 13:18:58.152001 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8csmw" event={"ID":"75a175f5-76d5-4a9e-920b-2b092642b9d1","Type":"ContainerStarted","Data":"333408ad9464c226acc3ca7c004d7e02d1d17edd73f2b9323b54c7c0a937b5fe"} Dec 04 13:18:59 crc kubenswrapper[4979]: I1204 13:18:59.164123 4979 generic.go:334] "Generic (PLEG): container finished" podID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerID="4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f" exitCode=0 Dec 04 13:18:59 crc kubenswrapper[4979]: I1204 13:18:59.164315 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8csmw" event={"ID":"75a175f5-76d5-4a9e-920b-2b092642b9d1","Type":"ContainerDied","Data":"4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f"} Dec 04 13:19:00 crc kubenswrapper[4979]: I1204 13:19:00.180667 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8csmw" event={"ID":"75a175f5-76d5-4a9e-920b-2b092642b9d1","Type":"ContainerStarted","Data":"fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433"} Dec 04 13:19:00 crc kubenswrapper[4979]: I1204 13:19:00.222922 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8csmw" podStartSLOduration=2.723020349 podStartE2EDuration="4.222898271s" podCreationTimestamp="2025-12-04 13:18:56 +0000 UTC" firstStartedPulling="2025-12-04 13:18:58.154241099 +0000 UTC m=+5762.428536913" lastFinishedPulling="2025-12-04 13:18:59.654119021 +0000 UTC m=+5763.928414835" observedRunningTime="2025-12-04 13:19:00.213404479 +0000 UTC m=+5764.487700293" watchObservedRunningTime="2025-12-04 13:19:00.222898271 +0000 UTC m=+5764.497194075" Dec 04 13:19:03 crc kubenswrapper[4979]: I1204 13:19:03.199709 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:19:03 crc kubenswrapper[4979]: E1204 13:19:03.200816 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:19:06 crc kubenswrapper[4979]: I1204 13:19:06.907122 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:19:06 crc kubenswrapper[4979]: I1204 13:19:06.907745 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:19:06 crc kubenswrapper[4979]: I1204 13:19:06.958098 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:19:07 crc kubenswrapper[4979]: I1204 13:19:07.291572 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:19:07 crc kubenswrapper[4979]: I1204 13:19:07.345252 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8csmw"] Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.265433 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8csmw" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="registry-server" containerID="cri-o://fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433" gracePeriod=2 Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.762135 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.917237 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kprn8\" (UniqueName: \"kubernetes.io/projected/75a175f5-76d5-4a9e-920b-2b092642b9d1-kube-api-access-kprn8\") pod \"75a175f5-76d5-4a9e-920b-2b092642b9d1\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.917367 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-utilities\") pod \"75a175f5-76d5-4a9e-920b-2b092642b9d1\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.917489 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-catalog-content\") pod \"75a175f5-76d5-4a9e-920b-2b092642b9d1\" (UID: \"75a175f5-76d5-4a9e-920b-2b092642b9d1\") " Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.917994 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-utilities" (OuterVolumeSpecName: "utilities") pod "75a175f5-76d5-4a9e-920b-2b092642b9d1" (UID: "75a175f5-76d5-4a9e-920b-2b092642b9d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.922136 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75a175f5-76d5-4a9e-920b-2b092642b9d1-kube-api-access-kprn8" (OuterVolumeSpecName: "kube-api-access-kprn8") pod "75a175f5-76d5-4a9e-920b-2b092642b9d1" (UID: "75a175f5-76d5-4a9e-920b-2b092642b9d1"). InnerVolumeSpecName "kube-api-access-kprn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:19:09 crc kubenswrapper[4979]: I1204 13:19:09.960024 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "75a175f5-76d5-4a9e-920b-2b092642b9d1" (UID: "75a175f5-76d5-4a9e-920b-2b092642b9d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.020004 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.020040 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kprn8\" (UniqueName: \"kubernetes.io/projected/75a175f5-76d5-4a9e-920b-2b092642b9d1-kube-api-access-kprn8\") on node \"crc\" DevicePath \"\"" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.020053 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/75a175f5-76d5-4a9e-920b-2b092642b9d1-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.278253 4979 generic.go:334] "Generic (PLEG): container finished" podID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerID="fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433" exitCode=0 Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.278407 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8csmw" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.278433 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8csmw" event={"ID":"75a175f5-76d5-4a9e-920b-2b092642b9d1","Type":"ContainerDied","Data":"fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433"} Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.278559 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8csmw" event={"ID":"75a175f5-76d5-4a9e-920b-2b092642b9d1","Type":"ContainerDied","Data":"333408ad9464c226acc3ca7c004d7e02d1d17edd73f2b9323b54c7c0a937b5fe"} Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.278607 4979 scope.go:117] "RemoveContainer" containerID="fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.302260 4979 scope.go:117] "RemoveContainer" containerID="4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.309110 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8csmw"] Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.318771 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8csmw"] Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.326613 4979 scope.go:117] "RemoveContainer" containerID="8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.366990 4979 scope.go:117] "RemoveContainer" containerID="fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433" Dec 04 13:19:10 crc kubenswrapper[4979]: E1204 13:19:10.367423 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433\": container with ID starting with fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433 not found: ID does not exist" containerID="fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.367456 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433"} err="failed to get container status \"fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433\": rpc error: code = NotFound desc = could not find container \"fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433\": container with ID starting with fa8a8aa943ae751e5570d18c463e10bf25bfc6ef13b73237da60c0f86ad94433 not found: ID does not exist" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.367475 4979 scope.go:117] "RemoveContainer" containerID="4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f" Dec 04 13:19:10 crc kubenswrapper[4979]: E1204 13:19:10.367875 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f\": container with ID starting with 4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f not found: ID does not exist" containerID="4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.367933 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f"} err="failed to get container status \"4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f\": rpc error: code = NotFound desc = could not find container \"4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f\": container with ID starting with 4f823cd458acde543eb36fb14b37035f2a96be59ff7d29d6de217c36c925882f not found: ID does not exist" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.367969 4979 scope.go:117] "RemoveContainer" containerID="8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098" Dec 04 13:19:10 crc kubenswrapper[4979]: E1204 13:19:10.368269 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098\": container with ID starting with 8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098 not found: ID does not exist" containerID="8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098" Dec 04 13:19:10 crc kubenswrapper[4979]: I1204 13:19:10.368295 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098"} err="failed to get container status \"8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098\": rpc error: code = NotFound desc = could not find container \"8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098\": container with ID starting with 8a735913cc7623d8067a2d6e87eb4e7bc1d793a8599efb1623986f66e17ac098 not found: ID does not exist" Dec 04 13:19:12 crc kubenswrapper[4979]: I1204 13:19:12.210985 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" path="/var/lib/kubelet/pods/75a175f5-76d5-4a9e-920b-2b092642b9d1/volumes" Dec 04 13:19:14 crc kubenswrapper[4979]: I1204 13:19:14.200708 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:19:14 crc kubenswrapper[4979]: E1204 13:19:14.201348 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:19:29 crc kubenswrapper[4979]: I1204 13:19:29.199132 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:19:29 crc kubenswrapper[4979]: E1204 13:19:29.199912 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:19:42 crc kubenswrapper[4979]: I1204 13:19:42.199414 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:19:42 crc kubenswrapper[4979]: E1204 13:19:42.200240 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:19:55 crc kubenswrapper[4979]: I1204 13:19:55.200035 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:19:55 crc kubenswrapper[4979]: E1204 13:19:55.201408 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:20:06 crc kubenswrapper[4979]: I1204 13:20:06.206356 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:20:06 crc kubenswrapper[4979]: E1204 13:20:06.207163 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:20:07 crc kubenswrapper[4979]: I1204 13:20:07.401792 4979 scope.go:117] "RemoveContainer" containerID="7064714334f098d81b367567cebbb47b023b8e2ff638bdaf9537aa2d9dddfaa3" Dec 04 13:20:07 crc kubenswrapper[4979]: I1204 13:20:07.425949 4979 scope.go:117] "RemoveContainer" containerID="149fcdd085c77a149377a9dfeb600153259be2f67df4f56c2c405b353a2f8c57" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.150030 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nr9fp"] Dec 04 13:20:15 crc kubenswrapper[4979]: E1204 13:20:15.154479 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="registry-server" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.154608 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="registry-server" Dec 04 13:20:15 crc kubenswrapper[4979]: E1204 13:20:15.154719 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="extract-utilities" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.154803 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="extract-utilities" Dec 04 13:20:15 crc kubenswrapper[4979]: E1204 13:20:15.154901 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="extract-content" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.155057 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="extract-content" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.155541 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a175f5-76d5-4a9e-920b-2b092642b9d1" containerName="registry-server" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.158133 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.168982 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fp"] Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.238684 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-catalog-content\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.238804 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgz9f\" (UniqueName: \"kubernetes.io/projected/d1d950a2-4455-4935-b4c9-5aaf418e42fa-kube-api-access-pgz9f\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.238976 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-utilities\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.341200 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-catalog-content\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.341428 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgz9f\" (UniqueName: \"kubernetes.io/projected/d1d950a2-4455-4935-b4c9-5aaf418e42fa-kube-api-access-pgz9f\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.341568 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-utilities\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.341798 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-catalog-content\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.342132 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-utilities\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.363706 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgz9f\" (UniqueName: \"kubernetes.io/projected/d1d950a2-4455-4935-b4c9-5aaf418e42fa-kube-api-access-pgz9f\") pod \"redhat-operators-nr9fp\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.487712 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:15 crc kubenswrapper[4979]: I1204 13:20:15.988629 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fp"] Dec 04 13:20:16 crc kubenswrapper[4979]: I1204 13:20:16.846173 4979 generic.go:334] "Generic (PLEG): container finished" podID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerID="131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4" exitCode=0 Dec 04 13:20:16 crc kubenswrapper[4979]: I1204 13:20:16.846290 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerDied","Data":"131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4"} Dec 04 13:20:16 crc kubenswrapper[4979]: I1204 13:20:16.846866 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerStarted","Data":"e941ff95958f00ed2d10c0a8625cad25979076b28502245c59574b33d7b10ae9"} Dec 04 13:20:17 crc kubenswrapper[4979]: I1204 13:20:17.199088 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:20:17 crc kubenswrapper[4979]: E1204 13:20:17.199407 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:20:17 crc kubenswrapper[4979]: I1204 13:20:17.860641 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerStarted","Data":"61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b"} Dec 04 13:20:20 crc kubenswrapper[4979]: I1204 13:20:20.891708 4979 generic.go:334] "Generic (PLEG): container finished" podID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerID="61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b" exitCode=0 Dec 04 13:20:20 crc kubenswrapper[4979]: I1204 13:20:20.891794 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerDied","Data":"61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b"} Dec 04 13:20:21 crc kubenswrapper[4979]: I1204 13:20:21.904836 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerStarted","Data":"4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1"} Dec 04 13:20:21 crc kubenswrapper[4979]: I1204 13:20:21.929841 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nr9fp" podStartSLOduration=2.361999623 podStartE2EDuration="6.929823672s" podCreationTimestamp="2025-12-04 13:20:15 +0000 UTC" firstStartedPulling="2025-12-04 13:20:16.848383049 +0000 UTC m=+5841.122678853" lastFinishedPulling="2025-12-04 13:20:21.416207098 +0000 UTC m=+5845.690502902" observedRunningTime="2025-12-04 13:20:21.925343663 +0000 UTC m=+5846.199639487" watchObservedRunningTime="2025-12-04 13:20:21.929823672 +0000 UTC m=+5846.204119476" Dec 04 13:20:25 crc kubenswrapper[4979]: I1204 13:20:25.487964 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:25 crc kubenswrapper[4979]: I1204 13:20:25.488795 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:26 crc kubenswrapper[4979]: I1204 13:20:26.541839 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nr9fp" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="registry-server" probeResult="failure" output=< Dec 04 13:20:26 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 13:20:26 crc kubenswrapper[4979]: > Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.543665 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cv4bf"] Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.548867 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.552818 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-7vxvq" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.553066 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.555988 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cv4bf"] Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.568270 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-dfk2m"] Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.570354 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.579475 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dfk2m"] Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.590783 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkpb5\" (UniqueName: \"kubernetes.io/projected/75e9db6b-7c43-497c-a993-34ea88e91b8f-kube-api-access-jkpb5\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.590850 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848fa47f-ef4c-4eef-9a0f-e0536131bba3-scripts\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.590873 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-lib\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.590893 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-run\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.590931 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-log\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.590968 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75e9db6b-7c43-497c-a993-34ea88e91b8f-scripts\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.591000 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-run\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.591034 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-run-ovn\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.591050 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-log-ovn\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.591075 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-etc-ovs\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.591093 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4gh2\" (UniqueName: \"kubernetes.io/projected/848fa47f-ef4c-4eef-9a0f-e0536131bba3-kube-api-access-h4gh2\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693093 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848fa47f-ef4c-4eef-9a0f-e0536131bba3-scripts\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693147 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-lib\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693174 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-run\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693228 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-log\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693325 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75e9db6b-7c43-497c-a993-34ea88e91b8f-scripts\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693371 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-run\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693416 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-run-ovn\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-log-ovn\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693473 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-etc-ovs\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693501 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4gh2\" (UniqueName: \"kubernetes.io/projected/848fa47f-ef4c-4eef-9a0f-e0536131bba3-kube-api-access-h4gh2\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693546 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkpb5\" (UniqueName: \"kubernetes.io/projected/75e9db6b-7c43-497c-a993-34ea88e91b8f-kube-api-access-jkpb5\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693767 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-log\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693806 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-lib\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693848 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-run\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693846 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-run-ovn\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693900 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-etc-ovs\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693960 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/75e9db6b-7c43-497c-a993-34ea88e91b8f-var-log-ovn\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.693958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/848fa47f-ef4c-4eef-9a0f-e0536131bba3-var-run\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.695406 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/848fa47f-ef4c-4eef-9a0f-e0536131bba3-scripts\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.695759 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75e9db6b-7c43-497c-a993-34ea88e91b8f-scripts\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.712031 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4gh2\" (UniqueName: \"kubernetes.io/projected/848fa47f-ef4c-4eef-9a0f-e0536131bba3-kube-api-access-h4gh2\") pod \"ovn-controller-ovs-dfk2m\" (UID: \"848fa47f-ef4c-4eef-9a0f-e0536131bba3\") " pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.715482 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkpb5\" (UniqueName: \"kubernetes.io/projected/75e9db6b-7c43-497c-a993-34ea88e91b8f-kube-api-access-jkpb5\") pod \"ovn-controller-cv4bf\" (UID: \"75e9db6b-7c43-497c-a993-34ea88e91b8f\") " pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.868757 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:28 crc kubenswrapper[4979]: I1204 13:20:28.907023 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:29 crc kubenswrapper[4979]: I1204 13:20:29.200659 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:20:29 crc kubenswrapper[4979]: E1204 13:20:29.202010 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:20:29 crc kubenswrapper[4979]: I1204 13:20:29.438815 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cv4bf"] Dec 04 13:20:29 crc kubenswrapper[4979]: W1204 13:20:29.870745 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod848fa47f_ef4c_4eef_9a0f_e0536131bba3.slice/crio-b960243311725fa0707de78cc5041bd40ac9458e9b9d3280bb25d590b9cbb8fa WatchSource:0}: Error finding container b960243311725fa0707de78cc5041bd40ac9458e9b9d3280bb25d590b9cbb8fa: Status 404 returned error can't find the container with id b960243311725fa0707de78cc5041bd40ac9458e9b9d3280bb25d590b9cbb8fa Dec 04 13:20:29 crc kubenswrapper[4979]: I1204 13:20:29.872275 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dfk2m"] Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.004498 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf" event={"ID":"75e9db6b-7c43-497c-a993-34ea88e91b8f","Type":"ContainerStarted","Data":"39600ff9f087e29618cd87841fc54a9fd6ff3a8636d6d22e9a33b3add5fdfa51"} Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.004871 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-cv4bf" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.004914 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf" event={"ID":"75e9db6b-7c43-497c-a993-34ea88e91b8f","Type":"ContainerStarted","Data":"3ffea40fcf52cd2bf1756ea5313c48b604b2ee12c4400c1befc90a2b3ab72e12"} Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.006715 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dfk2m" event={"ID":"848fa47f-ef4c-4eef-9a0f-e0536131bba3","Type":"ContainerStarted","Data":"b960243311725fa0707de78cc5041bd40ac9458e9b9d3280bb25d590b9cbb8fa"} Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.031488 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cv4bf" podStartSLOduration=2.031465614 podStartE2EDuration="2.031465614s" podCreationTimestamp="2025-12-04 13:20:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:20:30.024325024 +0000 UTC m=+5854.298620828" watchObservedRunningTime="2025-12-04 13:20:30.031465614 +0000 UTC m=+5854.305761418" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.184673 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-4kwz7"] Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.187094 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.189808 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.195090 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4kwz7"] Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.334539 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-config\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.334800 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-ovn-rundir\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.334850 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd8r4\" (UniqueName: \"kubernetes.io/projected/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-kube-api-access-zd8r4\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.334896 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-ovs-rundir\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.437514 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-ovn-rundir\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.437591 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd8r4\" (UniqueName: \"kubernetes.io/projected/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-kube-api-access-zd8r4\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.437636 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-ovs-rundir\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.437704 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-config\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.438004 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-ovs-rundir\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.438024 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-ovn-rundir\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.438642 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-config\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.464081 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd8r4\" (UniqueName: \"kubernetes.io/projected/cd2b947c-fa08-4f71-88ed-cd2cc216cc1e-kube-api-access-zd8r4\") pod \"ovn-controller-metrics-4kwz7\" (UID: \"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e\") " pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.510660 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-4kwz7" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.582011 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-create-flkrp"] Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.583366 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.595529 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-flkrp"] Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.742837 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr2c7\" (UniqueName: \"kubernetes.io/projected/ac4aba8f-0d93-48bc-b772-2a09e06ac886-kube-api-access-sr2c7\") pod \"octavia-db-create-flkrp\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.743215 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac4aba8f-0d93-48bc-b772-2a09e06ac886-operator-scripts\") pod \"octavia-db-create-flkrp\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.846565 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr2c7\" (UniqueName: \"kubernetes.io/projected/ac4aba8f-0d93-48bc-b772-2a09e06ac886-kube-api-access-sr2c7\") pod \"octavia-db-create-flkrp\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.846661 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac4aba8f-0d93-48bc-b772-2a09e06ac886-operator-scripts\") pod \"octavia-db-create-flkrp\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.847682 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac4aba8f-0d93-48bc-b772-2a09e06ac886-operator-scripts\") pod \"octavia-db-create-flkrp\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.873160 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr2c7\" (UniqueName: \"kubernetes.io/projected/ac4aba8f-0d93-48bc-b772-2a09e06ac886-kube-api-access-sr2c7\") pod \"octavia-db-create-flkrp\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:30 crc kubenswrapper[4979]: I1204 13:20:30.917564 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.004440 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-4kwz7"] Dec 04 13:20:31 crc kubenswrapper[4979]: W1204 13:20:31.009578 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd2b947c_fa08_4f71_88ed_cd2cc216cc1e.slice/crio-dd96de985ebd3d9cb244c073bb09732dcf2a2f9545f68e66063e917f74bd3e26 WatchSource:0}: Error finding container dd96de985ebd3d9cb244c073bb09732dcf2a2f9545f68e66063e917f74bd3e26: Status 404 returned error can't find the container with id dd96de985ebd3d9cb244c073bb09732dcf2a2f9545f68e66063e917f74bd3e26 Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.020731 4979 generic.go:334] "Generic (PLEG): container finished" podID="848fa47f-ef4c-4eef-9a0f-e0536131bba3" containerID="1601bac74a47509531ee7736deb8db8d30461dba03d8b66566261cd23ce2d348" exitCode=0 Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.021805 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dfk2m" event={"ID":"848fa47f-ef4c-4eef-9a0f-e0536131bba3","Type":"ContainerDied","Data":"1601bac74a47509531ee7736deb8db8d30461dba03d8b66566261cd23ce2d348"} Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.391632 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-create-flkrp"] Dec 04 13:20:31 crc kubenswrapper[4979]: W1204 13:20:31.409287 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac4aba8f_0d93_48bc_b772_2a09e06ac886.slice/crio-44959bcf51ee62bdeb7d93fe79f0878d4fb92b3d7bb747800fcc3d0fd5fe37dc WatchSource:0}: Error finding container 44959bcf51ee62bdeb7d93fe79f0878d4fb92b3d7bb747800fcc3d0fd5fe37dc: Status 404 returned error can't find the container with id 44959bcf51ee62bdeb7d93fe79f0878d4fb92b3d7bb747800fcc3d0fd5fe37dc Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.867074 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-e8f1-account-create-update-s56mn"] Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.869104 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.881103 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-db-secret" Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.882621 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-e8f1-account-create-update-s56mn"] Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.969682 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43df5618-f033-4aa9-9076-9f9a4dd1363b-operator-scripts\") pod \"octavia-e8f1-account-create-update-s56mn\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:31 crc kubenswrapper[4979]: I1204 13:20:31.970083 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jt9n\" (UniqueName: \"kubernetes.io/projected/43df5618-f033-4aa9-9076-9f9a4dd1363b-kube-api-access-7jt9n\") pod \"octavia-e8f1-account-create-update-s56mn\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.037445 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dfk2m" event={"ID":"848fa47f-ef4c-4eef-9a0f-e0536131bba3","Type":"ContainerStarted","Data":"16328aa37912964c6c10a8af3d478ed38a62f0a44592c4e0f07d34e17c20febc"} Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.037507 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dfk2m" event={"ID":"848fa47f-ef4c-4eef-9a0f-e0536131bba3","Type":"ContainerStarted","Data":"f2f11a983a1a84735a52c2bd1f9c134e13a658a13c83e6301b044c819b272ab4"} Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.039535 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.039604 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.039683 4979 generic.go:334] "Generic (PLEG): container finished" podID="ac4aba8f-0d93-48bc-b772-2a09e06ac886" containerID="d5512089d199aeeeb850b0c3f379fc454214bbc3f9b317f3ccee3b1a7cf68bd4" exitCode=0 Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.039767 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-flkrp" event={"ID":"ac4aba8f-0d93-48bc-b772-2a09e06ac886","Type":"ContainerDied","Data":"d5512089d199aeeeb850b0c3f379fc454214bbc3f9b317f3ccee3b1a7cf68bd4"} Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.039794 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-flkrp" event={"ID":"ac4aba8f-0d93-48bc-b772-2a09e06ac886","Type":"ContainerStarted","Data":"44959bcf51ee62bdeb7d93fe79f0878d4fb92b3d7bb747800fcc3d0fd5fe37dc"} Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.042825 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4kwz7" event={"ID":"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e","Type":"ContainerStarted","Data":"c6c19b0d62bce90dbde4ec4295bbc232266956cf0346c154d57faf79bdc02cae"} Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.042902 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-4kwz7" event={"ID":"cd2b947c-fa08-4f71-88ed-cd2cc216cc1e","Type":"ContainerStarted","Data":"dd96de985ebd3d9cb244c073bb09732dcf2a2f9545f68e66063e917f74bd3e26"} Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.071064 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-dfk2m" podStartSLOduration=4.071036133 podStartE2EDuration="4.071036133s" podCreationTimestamp="2025-12-04 13:20:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:20:32.060916154 +0000 UTC m=+5856.335211958" watchObservedRunningTime="2025-12-04 13:20:32.071036133 +0000 UTC m=+5856.345331937" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.072005 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jt9n\" (UniqueName: \"kubernetes.io/projected/43df5618-f033-4aa9-9076-9f9a4dd1363b-kube-api-access-7jt9n\") pod \"octavia-e8f1-account-create-update-s56mn\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.072102 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43df5618-f033-4aa9-9076-9f9a4dd1363b-operator-scripts\") pod \"octavia-e8f1-account-create-update-s56mn\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.073106 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43df5618-f033-4aa9-9076-9f9a4dd1363b-operator-scripts\") pod \"octavia-e8f1-account-create-update-s56mn\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.089026 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-4kwz7" podStartSLOduration=2.089007121 podStartE2EDuration="2.089007121s" podCreationTimestamp="2025-12-04 13:20:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:20:32.078364049 +0000 UTC m=+5856.352659853" watchObservedRunningTime="2025-12-04 13:20:32.089007121 +0000 UTC m=+5856.363302925" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.094266 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jt9n\" (UniqueName: \"kubernetes.io/projected/43df5618-f033-4aa9-9076-9f9a4dd1363b-kube-api-access-7jt9n\") pod \"octavia-e8f1-account-create-update-s56mn\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.201062 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:32 crc kubenswrapper[4979]: I1204 13:20:32.648726 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-e8f1-account-create-update-s56mn"] Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.054552 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e8f1-account-create-update-s56mn" event={"ID":"43df5618-f033-4aa9-9076-9f9a4dd1363b","Type":"ContainerStarted","Data":"1a47303949bda25c349ddfd0c163c6692fb6d8b90efa78d1505a076be0f545e8"} Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.054942 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e8f1-account-create-update-s56mn" event={"ID":"43df5618-f033-4aa9-9076-9f9a4dd1363b","Type":"ContainerStarted","Data":"670c3679a317d120c865dda6c85d566ad20b22e614c92922c28aff692ebd35e2"} Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.077608 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-e8f1-account-create-update-s56mn" podStartSLOduration=2.077588541 podStartE2EDuration="2.077588541s" podCreationTimestamp="2025-12-04 13:20:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:20:33.068047687 +0000 UTC m=+5857.342343511" watchObservedRunningTime="2025-12-04 13:20:33.077588541 +0000 UTC m=+5857.351884345" Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.346171 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.501391 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac4aba8f-0d93-48bc-b772-2a09e06ac886-operator-scripts\") pod \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.501450 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr2c7\" (UniqueName: \"kubernetes.io/projected/ac4aba8f-0d93-48bc-b772-2a09e06ac886-kube-api-access-sr2c7\") pod \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\" (UID: \"ac4aba8f-0d93-48bc-b772-2a09e06ac886\") " Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.502532 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac4aba8f-0d93-48bc-b772-2a09e06ac886-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ac4aba8f-0d93-48bc-b772-2a09e06ac886" (UID: "ac4aba8f-0d93-48bc-b772-2a09e06ac886"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.511437 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac4aba8f-0d93-48bc-b772-2a09e06ac886-kube-api-access-sr2c7" (OuterVolumeSpecName: "kube-api-access-sr2c7") pod "ac4aba8f-0d93-48bc-b772-2a09e06ac886" (UID: "ac4aba8f-0d93-48bc-b772-2a09e06ac886"). InnerVolumeSpecName "kube-api-access-sr2c7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.604084 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ac4aba8f-0d93-48bc-b772-2a09e06ac886-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:33 crc kubenswrapper[4979]: I1204 13:20:33.604414 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr2c7\" (UniqueName: \"kubernetes.io/projected/ac4aba8f-0d93-48bc-b772-2a09e06ac886-kube-api-access-sr2c7\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:34 crc kubenswrapper[4979]: I1204 13:20:34.068563 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-create-flkrp" Dec 04 13:20:34 crc kubenswrapper[4979]: I1204 13:20:34.068571 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-create-flkrp" event={"ID":"ac4aba8f-0d93-48bc-b772-2a09e06ac886","Type":"ContainerDied","Data":"44959bcf51ee62bdeb7d93fe79f0878d4fb92b3d7bb747800fcc3d0fd5fe37dc"} Dec 04 13:20:34 crc kubenswrapper[4979]: I1204 13:20:34.068647 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44959bcf51ee62bdeb7d93fe79f0878d4fb92b3d7bb747800fcc3d0fd5fe37dc" Dec 04 13:20:34 crc kubenswrapper[4979]: I1204 13:20:34.071244 4979 generic.go:334] "Generic (PLEG): container finished" podID="43df5618-f033-4aa9-9076-9f9a4dd1363b" containerID="1a47303949bda25c349ddfd0c163c6692fb6d8b90efa78d1505a076be0f545e8" exitCode=0 Dec 04 13:20:34 crc kubenswrapper[4979]: I1204 13:20:34.071334 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e8f1-account-create-update-s56mn" event={"ID":"43df5618-f033-4aa9-9076-9f9a4dd1363b","Type":"ContainerDied","Data":"1a47303949bda25c349ddfd0c163c6692fb6d8b90efa78d1505a076be0f545e8"} Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.448667 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.534538 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.584325 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.640829 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43df5618-f033-4aa9-9076-9f9a4dd1363b-operator-scripts\") pod \"43df5618-f033-4aa9-9076-9f9a4dd1363b\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.640923 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jt9n\" (UniqueName: \"kubernetes.io/projected/43df5618-f033-4aa9-9076-9f9a4dd1363b-kube-api-access-7jt9n\") pod \"43df5618-f033-4aa9-9076-9f9a4dd1363b\" (UID: \"43df5618-f033-4aa9-9076-9f9a4dd1363b\") " Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.641567 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43df5618-f033-4aa9-9076-9f9a4dd1363b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "43df5618-f033-4aa9-9076-9f9a4dd1363b" (UID: "43df5618-f033-4aa9-9076-9f9a4dd1363b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.651996 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43df5618-f033-4aa9-9076-9f9a4dd1363b-kube-api-access-7jt9n" (OuterVolumeSpecName: "kube-api-access-7jt9n") pod "43df5618-f033-4aa9-9076-9f9a4dd1363b" (UID: "43df5618-f033-4aa9-9076-9f9a4dd1363b"). InnerVolumeSpecName "kube-api-access-7jt9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.743052 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jt9n\" (UniqueName: \"kubernetes.io/projected/43df5618-f033-4aa9-9076-9f9a4dd1363b-kube-api-access-7jt9n\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.743106 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/43df5618-f033-4aa9-9076-9f9a4dd1363b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:35 crc kubenswrapper[4979]: I1204 13:20:35.772243 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fp"] Dec 04 13:20:36 crc kubenswrapper[4979]: I1204 13:20:36.093617 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-e8f1-account-create-update-s56mn" event={"ID":"43df5618-f033-4aa9-9076-9f9a4dd1363b","Type":"ContainerDied","Data":"670c3679a317d120c865dda6c85d566ad20b22e614c92922c28aff692ebd35e2"} Dec 04 13:20:36 crc kubenswrapper[4979]: I1204 13:20:36.093993 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="670c3679a317d120c865dda6c85d566ad20b22e614c92922c28aff692ebd35e2" Dec 04 13:20:36 crc kubenswrapper[4979]: I1204 13:20:36.093665 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-e8f1-account-create-update-s56mn" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.102211 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nr9fp" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="registry-server" containerID="cri-o://4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1" gracePeriod=2 Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.522991 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-persistence-db-create-stmlb"] Dec 04 13:20:37 crc kubenswrapper[4979]: E1204 13:20:37.524103 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac4aba8f-0d93-48bc-b772-2a09e06ac886" containerName="mariadb-database-create" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.524210 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac4aba8f-0d93-48bc-b772-2a09e06ac886" containerName="mariadb-database-create" Dec 04 13:20:37 crc kubenswrapper[4979]: E1204 13:20:37.524316 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43df5618-f033-4aa9-9076-9f9a4dd1363b" containerName="mariadb-account-create-update" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.524392 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="43df5618-f033-4aa9-9076-9f9a4dd1363b" containerName="mariadb-account-create-update" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.524696 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="43df5618-f033-4aa9-9076-9f9a4dd1363b" containerName="mariadb-account-create-update" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.524795 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac4aba8f-0d93-48bc-b772-2a09e06ac886" containerName="mariadb-database-create" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.525749 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.531873 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-stmlb"] Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.680921 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a03b33-a912-4006-9247-24dec637cf71-operator-scripts\") pod \"octavia-persistence-db-create-stmlb\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.681767 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ckwk\" (UniqueName: \"kubernetes.io/projected/38a03b33-a912-4006-9247-24dec637cf71-kube-api-access-2ckwk\") pod \"octavia-persistence-db-create-stmlb\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.784091 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a03b33-a912-4006-9247-24dec637cf71-operator-scripts\") pod \"octavia-persistence-db-create-stmlb\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.784160 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ckwk\" (UniqueName: \"kubernetes.io/projected/38a03b33-a912-4006-9247-24dec637cf71-kube-api-access-2ckwk\") pod \"octavia-persistence-db-create-stmlb\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.785336 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a03b33-a912-4006-9247-24dec637cf71-operator-scripts\") pod \"octavia-persistence-db-create-stmlb\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.804009 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ckwk\" (UniqueName: \"kubernetes.io/projected/38a03b33-a912-4006-9247-24dec637cf71-kube-api-access-2ckwk\") pod \"octavia-persistence-db-create-stmlb\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:37 crc kubenswrapper[4979]: I1204 13:20:37.845001 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.067233 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-b7de-account-create-update-s64mx"] Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.072396 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.075286 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-persistence-db-secret" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.080114 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-b7de-account-create-update-s64mx"] Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.093634 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.122639 4979 generic.go:334] "Generic (PLEG): container finished" podID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerID="4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1" exitCode=0 Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.122685 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerDied","Data":"4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1"} Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.122710 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nr9fp" event={"ID":"d1d950a2-4455-4935-b4c9-5aaf418e42fa","Type":"ContainerDied","Data":"e941ff95958f00ed2d10c0a8625cad25979076b28502245c59574b33d7b10ae9"} Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.122726 4979 scope.go:117] "RemoveContainer" containerID="4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.122850 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nr9fp" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.148799 4979 scope.go:117] "RemoveContainer" containerID="61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.170624 4979 scope.go:117] "RemoveContainer" containerID="131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.191989 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-catalog-content\") pod \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.192272 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-utilities\") pod \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.192391 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgz9f\" (UniqueName: \"kubernetes.io/projected/d1d950a2-4455-4935-b4c9-5aaf418e42fa-kube-api-access-pgz9f\") pod \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\" (UID: \"d1d950a2-4455-4935-b4c9-5aaf418e42fa\") " Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.192730 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5182298-e0b7-4a7e-bb57-eda965d9a041-operator-scripts\") pod \"octavia-b7de-account-create-update-s64mx\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.192867 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzcwh\" (UniqueName: \"kubernetes.io/projected/c5182298-e0b7-4a7e-bb57-eda965d9a041-kube-api-access-xzcwh\") pod \"octavia-b7de-account-create-update-s64mx\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.193369 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-utilities" (OuterVolumeSpecName: "utilities") pod "d1d950a2-4455-4935-b4c9-5aaf418e42fa" (UID: "d1d950a2-4455-4935-b4c9-5aaf418e42fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.193753 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.197319 4979 scope.go:117] "RemoveContainer" containerID="4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1" Dec 04 13:20:38 crc kubenswrapper[4979]: E1204 13:20:38.199534 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1\": container with ID starting with 4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1 not found: ID does not exist" containerID="4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.199595 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1d950a2-4455-4935-b4c9-5aaf418e42fa-kube-api-access-pgz9f" (OuterVolumeSpecName: "kube-api-access-pgz9f") pod "d1d950a2-4455-4935-b4c9-5aaf418e42fa" (UID: "d1d950a2-4455-4935-b4c9-5aaf418e42fa"). InnerVolumeSpecName "kube-api-access-pgz9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.199594 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1"} err="failed to get container status \"4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1\": rpc error: code = NotFound desc = could not find container \"4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1\": container with ID starting with 4cf03a3c57c65cbe736ee35b0833cbd978d2083380013f0421ab0c0ec74d6ce1 not found: ID does not exist" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.199633 4979 scope.go:117] "RemoveContainer" containerID="61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b" Dec 04 13:20:38 crc kubenswrapper[4979]: E1204 13:20:38.200128 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b\": container with ID starting with 61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b not found: ID does not exist" containerID="61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.200159 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b"} err="failed to get container status \"61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b\": rpc error: code = NotFound desc = could not find container \"61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b\": container with ID starting with 61f94af49416f2c30c2e797c1d4d78ad5d971638633a81356cd060531f44662b not found: ID does not exist" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.200185 4979 scope.go:117] "RemoveContainer" containerID="131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4" Dec 04 13:20:38 crc kubenswrapper[4979]: E1204 13:20:38.200487 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4\": container with ID starting with 131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4 not found: ID does not exist" containerID="131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.200518 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4"} err="failed to get container status \"131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4\": rpc error: code = NotFound desc = could not find container \"131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4\": container with ID starting with 131d043f26a58d72609e2c14ec9aba862ff8371f70a63c865f27b69c62abeab4 not found: ID does not exist" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.296087 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5182298-e0b7-4a7e-bb57-eda965d9a041-operator-scripts\") pod \"octavia-b7de-account-create-update-s64mx\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.296182 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzcwh\" (UniqueName: \"kubernetes.io/projected/c5182298-e0b7-4a7e-bb57-eda965d9a041-kube-api-access-xzcwh\") pod \"octavia-b7de-account-create-update-s64mx\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.296311 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgz9f\" (UniqueName: \"kubernetes.io/projected/d1d950a2-4455-4935-b4c9-5aaf418e42fa-kube-api-access-pgz9f\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.297138 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5182298-e0b7-4a7e-bb57-eda965d9a041-operator-scripts\") pod \"octavia-b7de-account-create-update-s64mx\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.307367 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d1d950a2-4455-4935-b4c9-5aaf418e42fa" (UID: "d1d950a2-4455-4935-b4c9-5aaf418e42fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.314052 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzcwh\" (UniqueName: \"kubernetes.io/projected/c5182298-e0b7-4a7e-bb57-eda965d9a041-kube-api-access-xzcwh\") pod \"octavia-b7de-account-create-update-s64mx\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.366059 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-persistence-db-create-stmlb"] Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.397952 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d950a2-4455-4935-b4c9-5aaf418e42fa-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.412260 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.579719 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fp"] Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.593106 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nr9fp"] Dec 04 13:20:38 crc kubenswrapper[4979]: W1204 13:20:38.871131 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc5182298_e0b7_4a7e_bb57_eda965d9a041.slice/crio-3cb863e6bade28b04cbd2abce1309ce821773765127d0f4a1a2677fb19e29410 WatchSource:0}: Error finding container 3cb863e6bade28b04cbd2abce1309ce821773765127d0f4a1a2677fb19e29410: Status 404 returned error can't find the container with id 3cb863e6bade28b04cbd2abce1309ce821773765127d0f4a1a2677fb19e29410 Dec 04 13:20:38 crc kubenswrapper[4979]: I1204 13:20:38.875786 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-b7de-account-create-update-s64mx"] Dec 04 13:20:39 crc kubenswrapper[4979]: I1204 13:20:39.132993 4979 generic.go:334] "Generic (PLEG): container finished" podID="38a03b33-a912-4006-9247-24dec637cf71" containerID="8060200e439848054f284863dd785c266481d069d02cba0e7562cafe93c5cf90" exitCode=0 Dec 04 13:20:39 crc kubenswrapper[4979]: I1204 13:20:39.133112 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-stmlb" event={"ID":"38a03b33-a912-4006-9247-24dec637cf71","Type":"ContainerDied","Data":"8060200e439848054f284863dd785c266481d069d02cba0e7562cafe93c5cf90"} Dec 04 13:20:39 crc kubenswrapper[4979]: I1204 13:20:39.133146 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-stmlb" event={"ID":"38a03b33-a912-4006-9247-24dec637cf71","Type":"ContainerStarted","Data":"c717d0ea424df4f5701ab1cfd763d5252d10d4614bdc63961ea7b5f647d8ec27"} Dec 04 13:20:39 crc kubenswrapper[4979]: I1204 13:20:39.138923 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-b7de-account-create-update-s64mx" event={"ID":"c5182298-e0b7-4a7e-bb57-eda965d9a041","Type":"ContainerStarted","Data":"3cb863e6bade28b04cbd2abce1309ce821773765127d0f4a1a2677fb19e29410"} Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.149358 4979 generic.go:334] "Generic (PLEG): container finished" podID="c5182298-e0b7-4a7e-bb57-eda965d9a041" containerID="5022f8314541adfdca2a69cdf8f131da56377026a2efb1db45231519a5f31d5c" exitCode=0 Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.149456 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-b7de-account-create-update-s64mx" event={"ID":"c5182298-e0b7-4a7e-bb57-eda965d9a041","Type":"ContainerDied","Data":"5022f8314541adfdca2a69cdf8f131da56377026a2efb1db45231519a5f31d5c"} Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.210658 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" path="/var/lib/kubelet/pods/d1d950a2-4455-4935-b4c9-5aaf418e42fa/volumes" Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.501424 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.640500 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a03b33-a912-4006-9247-24dec637cf71-operator-scripts\") pod \"38a03b33-a912-4006-9247-24dec637cf71\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.640571 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ckwk\" (UniqueName: \"kubernetes.io/projected/38a03b33-a912-4006-9247-24dec637cf71-kube-api-access-2ckwk\") pod \"38a03b33-a912-4006-9247-24dec637cf71\" (UID: \"38a03b33-a912-4006-9247-24dec637cf71\") " Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.641497 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/38a03b33-a912-4006-9247-24dec637cf71-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "38a03b33-a912-4006-9247-24dec637cf71" (UID: "38a03b33-a912-4006-9247-24dec637cf71"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.647130 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38a03b33-a912-4006-9247-24dec637cf71-kube-api-access-2ckwk" (OuterVolumeSpecName: "kube-api-access-2ckwk") pod "38a03b33-a912-4006-9247-24dec637cf71" (UID: "38a03b33-a912-4006-9247-24dec637cf71"). InnerVolumeSpecName "kube-api-access-2ckwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.743538 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ckwk\" (UniqueName: \"kubernetes.io/projected/38a03b33-a912-4006-9247-24dec637cf71-kube-api-access-2ckwk\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:40 crc kubenswrapper[4979]: I1204 13:20:40.743583 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/38a03b33-a912-4006-9247-24dec637cf71-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.053417 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-xckk8"] Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.062405 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-xckk8"] Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.165372 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-persistence-db-create-stmlb" event={"ID":"38a03b33-a912-4006-9247-24dec637cf71","Type":"ContainerDied","Data":"c717d0ea424df4f5701ab1cfd763d5252d10d4614bdc63961ea7b5f647d8ec27"} Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.165468 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c717d0ea424df4f5701ab1cfd763d5252d10d4614bdc63961ea7b5f647d8ec27" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.165400 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-persistence-db-create-stmlb" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.198790 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:20:41 crc kubenswrapper[4979]: E1204 13:20:41.199194 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.546764 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.662360 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5182298-e0b7-4a7e-bb57-eda965d9a041-operator-scripts\") pod \"c5182298-e0b7-4a7e-bb57-eda965d9a041\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.662618 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzcwh\" (UniqueName: \"kubernetes.io/projected/c5182298-e0b7-4a7e-bb57-eda965d9a041-kube-api-access-xzcwh\") pod \"c5182298-e0b7-4a7e-bb57-eda965d9a041\" (UID: \"c5182298-e0b7-4a7e-bb57-eda965d9a041\") " Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.663126 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5182298-e0b7-4a7e-bb57-eda965d9a041-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c5182298-e0b7-4a7e-bb57-eda965d9a041" (UID: "c5182298-e0b7-4a7e-bb57-eda965d9a041"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.666837 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5182298-e0b7-4a7e-bb57-eda965d9a041-kube-api-access-xzcwh" (OuterVolumeSpecName: "kube-api-access-xzcwh") pod "c5182298-e0b7-4a7e-bb57-eda965d9a041" (UID: "c5182298-e0b7-4a7e-bb57-eda965d9a041"). InnerVolumeSpecName "kube-api-access-xzcwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.764423 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzcwh\" (UniqueName: \"kubernetes.io/projected/c5182298-e0b7-4a7e-bb57-eda965d9a041-kube-api-access-xzcwh\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:41 crc kubenswrapper[4979]: I1204 13:20:41.764472 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c5182298-e0b7-4a7e-bb57-eda965d9a041-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.035788 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-0762-account-create-update-f26mh"] Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.044094 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-0762-account-create-update-f26mh"] Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.176467 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-b7de-account-create-update-s64mx" event={"ID":"c5182298-e0b7-4a7e-bb57-eda965d9a041","Type":"ContainerDied","Data":"3cb863e6bade28b04cbd2abce1309ce821773765127d0f4a1a2677fb19e29410"} Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.176509 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3cb863e6bade28b04cbd2abce1309ce821773765127d0f4a1a2677fb19e29410" Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.176562 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-b7de-account-create-update-s64mx" Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.223658 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="917d86c8-d0fd-4769-aa6f-f1bc490c1406" path="/var/lib/kubelet/pods/917d86c8-d0fd-4769-aa6f-f1bc490c1406/volumes" Dec 04 13:20:42 crc kubenswrapper[4979]: I1204 13:20:42.224513 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e57348b9-9bff-4b87-8075-b8bf9e459d02" path="/var/lib/kubelet/pods/e57348b9-9bff-4b87-8075-b8bf9e459d02/volumes" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.639461 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-api-7d99f6b98b-p9sbq"] Dec 04 13:20:43 crc kubenswrapper[4979]: E1204 13:20:43.639857 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5182298-e0b7-4a7e-bb57-eda965d9a041" containerName="mariadb-account-create-update" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.639869 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5182298-e0b7-4a7e-bb57-eda965d9a041" containerName="mariadb-account-create-update" Dec 04 13:20:43 crc kubenswrapper[4979]: E1204 13:20:43.639877 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38a03b33-a912-4006-9247-24dec637cf71" containerName="mariadb-database-create" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.639883 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="38a03b33-a912-4006-9247-24dec637cf71" containerName="mariadb-database-create" Dec 04 13:20:43 crc kubenswrapper[4979]: E1204 13:20:43.639894 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="extract-utilities" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.639900 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="extract-utilities" Dec 04 13:20:43 crc kubenswrapper[4979]: E1204 13:20:43.639913 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="extract-content" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.639918 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="extract-content" Dec 04 13:20:43 crc kubenswrapper[4979]: E1204 13:20:43.639939 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="registry-server" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.639945 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="registry-server" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.640133 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5182298-e0b7-4a7e-bb57-eda965d9a041" containerName="mariadb-account-create-update" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.640143 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="38a03b33-a912-4006-9247-24dec637cf71" containerName="mariadb-database-create" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.640153 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1d950a2-4455-4935-b4c9-5aaf418e42fa" containerName="registry-server" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.641583 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.648590 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-scripts" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.648652 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-octavia-dockercfg-7fd8w" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.649051 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-api-config-data" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.668126 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7d99f6b98b-p9sbq"] Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.807692 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-config-data\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.807966 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e3146a9d-fd80-475d-a9c2-cf52b42b437a-config-data-merged\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.808058 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-scripts\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.808136 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/e3146a9d-fd80-475d-a9c2-cf52b42b437a-octavia-run\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.808217 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-combined-ca-bundle\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.910429 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-config-data\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.910500 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e3146a9d-fd80-475d-a9c2-cf52b42b437a-config-data-merged\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.910532 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-scripts\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.910560 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/e3146a9d-fd80-475d-a9c2-cf52b42b437a-octavia-run\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.910592 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-combined-ca-bundle\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.911097 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"octavia-run\" (UniqueName: \"kubernetes.io/empty-dir/e3146a9d-fd80-475d-a9c2-cf52b42b437a-octavia-run\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.911138 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/e3146a9d-fd80-475d-a9c2-cf52b42b437a-config-data-merged\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.916395 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-combined-ca-bundle\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.916487 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-config-data\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.917210 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e3146a9d-fd80-475d-a9c2-cf52b42b437a-scripts\") pod \"octavia-api-7d99f6b98b-p9sbq\" (UID: \"e3146a9d-fd80-475d-a9c2-cf52b42b437a\") " pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:43 crc kubenswrapper[4979]: I1204 13:20:43.968208 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:44 crc kubenswrapper[4979]: I1204 13:20:44.589880 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-api-7d99f6b98b-p9sbq"] Dec 04 13:20:44 crc kubenswrapper[4979]: W1204 13:20:44.593531 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode3146a9d_fd80_475d_a9c2_cf52b42b437a.slice/crio-f27883b6d88043a0571c908e44635812838e8c726316a393f2c9ced3f14984a6 WatchSource:0}: Error finding container f27883b6d88043a0571c908e44635812838e8c726316a393f2c9ced3f14984a6: Status 404 returned error can't find the container with id f27883b6d88043a0571c908e44635812838e8c726316a393f2c9ced3f14984a6 Dec 04 13:20:45 crc kubenswrapper[4979]: I1204 13:20:45.201366 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7d99f6b98b-p9sbq" event={"ID":"e3146a9d-fd80-475d-a9c2-cf52b42b437a","Type":"ContainerStarted","Data":"f27883b6d88043a0571c908e44635812838e8c726316a393f2c9ced3f14984a6"} Dec 04 13:20:50 crc kubenswrapper[4979]: I1204 13:20:50.034982 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-pqvcx"] Dec 04 13:20:50 crc kubenswrapper[4979]: I1204 13:20:50.046230 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-pqvcx"] Dec 04 13:20:50 crc kubenswrapper[4979]: I1204 13:20:50.212608 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6753d199-6125-4b42-bfa8-cae894e9aa06" path="/var/lib/kubelet/pods/6753d199-6125-4b42-bfa8-cae894e9aa06/volumes" Dec 04 13:20:52 crc kubenswrapper[4979]: I1204 13:20:52.199121 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:20:52 crc kubenswrapper[4979]: E1204 13:20:52.199821 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:20:54 crc kubenswrapper[4979]: I1204 13:20:54.291969 4979 generic.go:334] "Generic (PLEG): container finished" podID="e3146a9d-fd80-475d-a9c2-cf52b42b437a" containerID="b6a2f01b79495c151c794ef69fa0fafad5016462cd58c30292e526b9471025a3" exitCode=0 Dec 04 13:20:54 crc kubenswrapper[4979]: I1204 13:20:54.292018 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7d99f6b98b-p9sbq" event={"ID":"e3146a9d-fd80-475d-a9c2-cf52b42b437a","Type":"ContainerDied","Data":"b6a2f01b79495c151c794ef69fa0fafad5016462cd58c30292e526b9471025a3"} Dec 04 13:20:55 crc kubenswrapper[4979]: I1204 13:20:55.304455 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7d99f6b98b-p9sbq" event={"ID":"e3146a9d-fd80-475d-a9c2-cf52b42b437a","Type":"ContainerStarted","Data":"9899d9670ea832ccd06669db9415afb2ae3122f85a06f9f57846b1bb2da44474"} Dec 04 13:20:55 crc kubenswrapper[4979]: I1204 13:20:55.305076 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-api-7d99f6b98b-p9sbq" event={"ID":"e3146a9d-fd80-475d-a9c2-cf52b42b437a","Type":"ContainerStarted","Data":"5c11d2edbd70d9632cc15b0b555ab199d6bef193d2d5a009641bc15974f97763"} Dec 04 13:20:55 crc kubenswrapper[4979]: I1204 13:20:55.306350 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:20:55 crc kubenswrapper[4979]: I1204 13:20:55.306389 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:21:03 crc kubenswrapper[4979]: I1204 13:21:03.903395 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-cv4bf" podUID="75e9db6b-7c43-497c-a993-34ea88e91b8f" containerName="ovn-controller" probeResult="failure" output=< Dec 04 13:21:03 crc kubenswrapper[4979]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 04 13:21:03 crc kubenswrapper[4979]: > Dec 04 13:21:03 crc kubenswrapper[4979]: I1204 13:21:03.951559 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:21:03 crc kubenswrapper[4979]: I1204 13:21:03.953954 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dfk2m" Dec 04 13:21:03 crc kubenswrapper[4979]: I1204 13:21:03.973807 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-api-7d99f6b98b-p9sbq" podStartSLOduration=11.794760225 podStartE2EDuration="20.973787236s" podCreationTimestamp="2025-12-04 13:20:43 +0000 UTC" firstStartedPulling="2025-12-04 13:20:44.595830229 +0000 UTC m=+5868.870126033" lastFinishedPulling="2025-12-04 13:20:53.77485722 +0000 UTC m=+5878.049153044" observedRunningTime="2025-12-04 13:20:55.325735889 +0000 UTC m=+5879.600031693" watchObservedRunningTime="2025-12-04 13:21:03.973787236 +0000 UTC m=+5888.248083040" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.090952 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cv4bf-config-j4slc"] Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.092436 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.095924 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.104767 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cv4bf-config-j4slc"] Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.201389 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-scripts\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.201707 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-additional-scripts\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.201731 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm74n\" (UniqueName: \"kubernetes.io/projected/630fbc08-052b-4209-b0c6-00b288bdd384-kube-api-access-fm74n\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.201766 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run-ovn\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.201795 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-log-ovn\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.201816 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.303458 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run-ovn\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.303559 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-log-ovn\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.303586 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.303833 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-scripts\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.303872 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-additional-scripts\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.303892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm74n\" (UniqueName: \"kubernetes.io/projected/630fbc08-052b-4209-b0c6-00b288bdd384-kube-api-access-fm74n\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.304355 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run-ovn\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.304457 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-log-ovn\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.304925 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.305958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-additional-scripts\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.307695 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-scripts\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.329606 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm74n\" (UniqueName: \"kubernetes.io/projected/630fbc08-052b-4209-b0c6-00b288bdd384-kube-api-access-fm74n\") pod \"ovn-controller-cv4bf-config-j4slc\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.424641 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:04 crc kubenswrapper[4979]: I1204 13:21:04.904475 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cv4bf-config-j4slc"] Dec 04 13:21:05 crc kubenswrapper[4979]: I1204 13:21:05.059826 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-958bg"] Dec 04 13:21:05 crc kubenswrapper[4979]: I1204 13:21:05.068459 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-958bg"] Dec 04 13:21:05 crc kubenswrapper[4979]: I1204 13:21:05.198931 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:21:05 crc kubenswrapper[4979]: E1204 13:21:05.199523 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:21:05 crc kubenswrapper[4979]: I1204 13:21:05.397229 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-j4slc" event={"ID":"630fbc08-052b-4209-b0c6-00b288bdd384","Type":"ContainerStarted","Data":"6ed5d177f39e6c21dd633f04a91a95b8540f7e762e4cd046d1ec30e32d0328a5"} Dec 04 13:21:06 crc kubenswrapper[4979]: I1204 13:21:06.210640 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa6e5e6-5309-489a-b5f7-1b6f34afac11" path="/var/lib/kubelet/pods/5fa6e5e6-5309-489a-b5f7-1b6f34afac11/volumes" Dec 04 13:21:06 crc kubenswrapper[4979]: I1204 13:21:06.411292 4979 generic.go:334] "Generic (PLEG): container finished" podID="630fbc08-052b-4209-b0c6-00b288bdd384" containerID="db71e79090ed41fbcf86fe025d7f936cca867a9118e67fc3b1d09dc2e175fa87" exitCode=0 Dec 04 13:21:06 crc kubenswrapper[4979]: I1204 13:21:06.411353 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-j4slc" event={"ID":"630fbc08-052b-4209-b0c6-00b288bdd384","Type":"ContainerDied","Data":"db71e79090ed41fbcf86fe025d7f936cca867a9118e67fc3b1d09dc2e175fa87"} Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.200605 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-rsyslog-5bm9b"] Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.202514 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.205098 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-config-data" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.205272 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"octavia-hmport-map" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.205939 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-rsyslog-scripts" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.211866 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-5bm9b"] Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.262340 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d361aab-dd58-4fc9-aa52-738ff495f640-config-data-merged\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.262504 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1d361aab-dd58-4fc9-aa52-738ff495f640-hm-ports\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.262545 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d361aab-dd58-4fc9-aa52-738ff495f640-config-data\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.262657 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d361aab-dd58-4fc9-aa52-738ff495f640-scripts\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.364056 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d361aab-dd58-4fc9-aa52-738ff495f640-scripts\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.364200 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d361aab-dd58-4fc9-aa52-738ff495f640-config-data-merged\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.364257 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1d361aab-dd58-4fc9-aa52-738ff495f640-hm-ports\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.364284 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d361aab-dd58-4fc9-aa52-738ff495f640-config-data\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.364719 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d361aab-dd58-4fc9-aa52-738ff495f640-config-data-merged\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.365547 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/1d361aab-dd58-4fc9-aa52-738ff495f640-hm-ports\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.371533 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d361aab-dd58-4fc9-aa52-738ff495f640-scripts\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.379039 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d361aab-dd58-4fc9-aa52-738ff495f640-config-data\") pod \"octavia-rsyslog-5bm9b\" (UID: \"1d361aab-dd58-4fc9-aa52-738ff495f640\") " pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.521570 4979 scope.go:117] "RemoveContainer" containerID="a4960334836f7e198fd176ea85315ee9f98088e691848d494e35542d764e7349" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.526355 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.590589 4979 scope.go:117] "RemoveContainer" containerID="a9d1bdb9a9b4e6fb7af3f9d4801cc3463ca8134a3a4ccb3089d5cd0579d83cb9" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.700199 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.702490 4979 scope.go:117] "RemoveContainer" containerID="0951ba9ea4751f2febbfaac4bae14ed3efd49fbbdbf2a38c901e3db830f44fcd" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.749952 4979 scope.go:117] "RemoveContainer" containerID="b554a368d45951b3c430a573c940dd834684b37e97d23fc863303c50f2b7e2da" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.779595 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-scripts\") pod \"630fbc08-052b-4209-b0c6-00b288bdd384\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.779668 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run-ovn\") pod \"630fbc08-052b-4209-b0c6-00b288bdd384\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.779717 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-log-ovn\") pod \"630fbc08-052b-4209-b0c6-00b288bdd384\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.779796 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run\") pod \"630fbc08-052b-4209-b0c6-00b288bdd384\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.779851 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm74n\" (UniqueName: \"kubernetes.io/projected/630fbc08-052b-4209-b0c6-00b288bdd384-kube-api-access-fm74n\") pod \"630fbc08-052b-4209-b0c6-00b288bdd384\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.779871 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-additional-scripts\") pod \"630fbc08-052b-4209-b0c6-00b288bdd384\" (UID: \"630fbc08-052b-4209-b0c6-00b288bdd384\") " Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.780828 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-scripts" (OuterVolumeSpecName: "scripts") pod "630fbc08-052b-4209-b0c6-00b288bdd384" (UID: "630fbc08-052b-4209-b0c6-00b288bdd384"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.780892 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "630fbc08-052b-4209-b0c6-00b288bdd384" (UID: "630fbc08-052b-4209-b0c6-00b288bdd384"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.780916 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "630fbc08-052b-4209-b0c6-00b288bdd384" (UID: "630fbc08-052b-4209-b0c6-00b288bdd384"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.780933 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run" (OuterVolumeSpecName: "var-run") pod "630fbc08-052b-4209-b0c6-00b288bdd384" (UID: "630fbc08-052b-4209-b0c6-00b288bdd384"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.781110 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "630fbc08-052b-4209-b0c6-00b288bdd384" (UID: "630fbc08-052b-4209-b0c6-00b288bdd384"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.787770 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/630fbc08-052b-4209-b0c6-00b288bdd384-kube-api-access-fm74n" (OuterVolumeSpecName: "kube-api-access-fm74n") pod "630fbc08-052b-4209-b0c6-00b288bdd384" (UID: "630fbc08-052b-4209-b0c6-00b288bdd384"). InnerVolumeSpecName "kube-api-access-fm74n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.882617 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.882903 4979 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.882916 4979 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.882927 4979 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/630fbc08-052b-4209-b0c6-00b288bdd384-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.882938 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm74n\" (UniqueName: \"kubernetes.io/projected/630fbc08-052b-4209-b0c6-00b288bdd384-kube-api-access-fm74n\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.882951 4979 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/630fbc08-052b-4209-b0c6-00b288bdd384-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.888310 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-x86s4"] Dec 04 13:21:07 crc kubenswrapper[4979]: E1204 13:21:07.888802 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="630fbc08-052b-4209-b0c6-00b288bdd384" containerName="ovn-config" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.888824 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="630fbc08-052b-4209-b0c6-00b288bdd384" containerName="ovn-config" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.889096 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="630fbc08-052b-4209-b0c6-00b288bdd384" containerName="ovn-config" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.890409 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.893919 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.903632 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-x86s4"] Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.984561 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/06bddff5-a672-4ac8-810d-3332a26a8344-httpd-config\") pod \"octavia-image-upload-59f8cff499-x86s4\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:07 crc kubenswrapper[4979]: I1204 13:21:07.984974 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/06bddff5-a672-4ac8-810d-3332a26a8344-amphora-image\") pod \"octavia-image-upload-59f8cff499-x86s4\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.086691 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/06bddff5-a672-4ac8-810d-3332a26a8344-amphora-image\") pod \"octavia-image-upload-59f8cff499-x86s4\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.086894 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/06bddff5-a672-4ac8-810d-3332a26a8344-httpd-config\") pod \"octavia-image-upload-59f8cff499-x86s4\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.087326 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/06bddff5-a672-4ac8-810d-3332a26a8344-amphora-image\") pod \"octavia-image-upload-59f8cff499-x86s4\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.097474 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/06bddff5-a672-4ac8-810d-3332a26a8344-httpd-config\") pod \"octavia-image-upload-59f8cff499-x86s4\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.105820 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-rsyslog-5bm9b"] Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.218584 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.433473 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-j4slc" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.433475 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-j4slc" event={"ID":"630fbc08-052b-4209-b0c6-00b288bdd384","Type":"ContainerDied","Data":"6ed5d177f39e6c21dd633f04a91a95b8540f7e762e4cd046d1ec30e32d0328a5"} Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.433621 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ed5d177f39e6c21dd633f04a91a95b8540f7e762e4cd046d1ec30e32d0328a5" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.435760 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-5bm9b" event={"ID":"1d361aab-dd58-4fc9-aa52-738ff495f640","Type":"ContainerStarted","Data":"50dd3882a277e935f48348d43ba8a23d0836c6efe541b164c18cb33916358604"} Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.704464 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-x86s4"] Dec 04 13:21:08 crc kubenswrapper[4979]: W1204 13:21:08.709023 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod06bddff5_a672_4ac8_810d_3332a26a8344.slice/crio-c906eb6793b80f3f6420b497a06e08985500bf6f6d6d4b21aeec6a408e0dcccf WatchSource:0}: Error finding container c906eb6793b80f3f6420b497a06e08985500bf6f6d6d4b21aeec6a408e0dcccf: Status 404 returned error can't find the container with id c906eb6793b80f3f6420b497a06e08985500bf6f6d6d4b21aeec6a408e0dcccf Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.789671 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cv4bf-config-j4slc"] Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.803064 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cv4bf-config-j4slc"] Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.915012 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-cv4bf-config-cknxb"] Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.916519 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.919723 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.932880 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cv4bf-config-cknxb"] Dec 04 13:21:08 crc kubenswrapper[4979]: I1204 13:21:08.935555 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-cv4bf" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.002232 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run-ovn\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.002356 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-additional-scripts\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.002462 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.002498 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-log-ovn\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.002548 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8fmr\" (UniqueName: \"kubernetes.io/projected/77b25d9f-8183-4c32-b52d-4f2933cd8471-kube-api-access-f8fmr\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.002605 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-scripts\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104093 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run-ovn\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104190 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-additional-scripts\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104295 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104345 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-log-ovn\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104390 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8fmr\" (UniqueName: \"kubernetes.io/projected/77b25d9f-8183-4c32-b52d-4f2933cd8471-kube-api-access-f8fmr\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104437 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-scripts\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104551 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104551 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run-ovn\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.104641 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-log-ovn\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.105345 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-additional-scripts\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.106682 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-scripts\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.141570 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8fmr\" (UniqueName: \"kubernetes.io/projected/77b25d9f-8183-4c32-b52d-4f2933cd8471-kube-api-access-f8fmr\") pod \"ovn-controller-cv4bf-config-cknxb\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.265799 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:09 crc kubenswrapper[4979]: I1204 13:21:09.445982 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-x86s4" event={"ID":"06bddff5-a672-4ac8-810d-3332a26a8344","Type":"ContainerStarted","Data":"c906eb6793b80f3f6420b497a06e08985500bf6f6d6d4b21aeec6a408e0dcccf"} Dec 04 13:21:10 crc kubenswrapper[4979]: I1204 13:21:10.223875 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="630fbc08-052b-4209-b0c6-00b288bdd384" path="/var/lib/kubelet/pods/630fbc08-052b-4209-b0c6-00b288bdd384/volumes" Dec 04 13:21:11 crc kubenswrapper[4979]: I1204 13:21:11.140169 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-cv4bf-config-cknxb"] Dec 04 13:21:11 crc kubenswrapper[4979]: I1204 13:21:11.472698 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-cknxb" event={"ID":"77b25d9f-8183-4c32-b52d-4f2933cd8471","Type":"ContainerStarted","Data":"db3ddc62c92913898afb5a210ad74285a94ad33a1681d44fc02ca4fe578971a5"} Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.450617 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-db-sync-pscsr"] Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.452999 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.455624 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-scripts" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.476207 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-pscsr"] Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.503820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-cknxb" event={"ID":"77b25d9f-8183-4c32-b52d-4f2933cd8471","Type":"ContainerStarted","Data":"083d4f5c7da426c8af9e3254c77469de6f04118b249b0abef98f11b3d015a153"} Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.510398 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-5bm9b" event={"ID":"1d361aab-dd58-4fc9-aa52-738ff495f640","Type":"ContainerStarted","Data":"c5eb9dce3a1b97cc511999d95247912239f7a9e702f91370cac8f4d9bb315ea1"} Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.529138 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.529224 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-scripts\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.529356 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data-merged\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.529431 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-combined-ca-bundle\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.529748 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-cv4bf-config-cknxb" podStartSLOduration=5.529727938 podStartE2EDuration="5.529727938s" podCreationTimestamp="2025-12-04 13:21:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:21:13.527852278 +0000 UTC m=+5897.802148082" watchObservedRunningTime="2025-12-04 13:21:13.529727938 +0000 UTC m=+5897.804023742" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.631286 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.631575 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-scripts\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.631754 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data-merged\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.631885 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-combined-ca-bundle\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.633337 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data-merged\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.638662 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-scripts\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.640152 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.640870 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-combined-ca-bundle\") pod \"octavia-db-sync-pscsr\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:13 crc kubenswrapper[4979]: I1204 13:21:13.782757 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:14 crc kubenswrapper[4979]: I1204 13:21:14.525666 4979 generic.go:334] "Generic (PLEG): container finished" podID="77b25d9f-8183-4c32-b52d-4f2933cd8471" containerID="083d4f5c7da426c8af9e3254c77469de6f04118b249b0abef98f11b3d015a153" exitCode=0 Dec 04 13:21:14 crc kubenswrapper[4979]: I1204 13:21:14.525823 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-cknxb" event={"ID":"77b25d9f-8183-4c32-b52d-4f2933cd8471","Type":"ContainerDied","Data":"083d4f5c7da426c8af9e3254c77469de6f04118b249b0abef98f11b3d015a153"} Dec 04 13:21:15 crc kubenswrapper[4979]: I1204 13:21:15.541083 4979 generic.go:334] "Generic (PLEG): container finished" podID="1d361aab-dd58-4fc9-aa52-738ff495f640" containerID="c5eb9dce3a1b97cc511999d95247912239f7a9e702f91370cac8f4d9bb315ea1" exitCode=0 Dec 04 13:21:15 crc kubenswrapper[4979]: I1204 13:21:15.541175 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-5bm9b" event={"ID":"1d361aab-dd58-4fc9-aa52-738ff495f640","Type":"ContainerDied","Data":"c5eb9dce3a1b97cc511999d95247912239f7a9e702f91370cac8f4d9bb315ea1"} Dec 04 13:21:18 crc kubenswrapper[4979]: I1204 13:21:18.479928 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:21:18 crc kubenswrapper[4979]: I1204 13:21:18.622758 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-api-7d99f6b98b-p9sbq" Dec 04 13:21:18 crc kubenswrapper[4979]: I1204 13:21:18.981208 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064205 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run-ovn\") pod \"77b25d9f-8183-4c32-b52d-4f2933cd8471\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064266 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run\") pod \"77b25d9f-8183-4c32-b52d-4f2933cd8471\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064323 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "77b25d9f-8183-4c32-b52d-4f2933cd8471" (UID: "77b25d9f-8183-4c32-b52d-4f2933cd8471"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064397 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-additional-scripts\") pod \"77b25d9f-8183-4c32-b52d-4f2933cd8471\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064447 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8fmr\" (UniqueName: \"kubernetes.io/projected/77b25d9f-8183-4c32-b52d-4f2933cd8471-kube-api-access-f8fmr\") pod \"77b25d9f-8183-4c32-b52d-4f2933cd8471\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064440 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run" (OuterVolumeSpecName: "var-run") pod "77b25d9f-8183-4c32-b52d-4f2933cd8471" (UID: "77b25d9f-8183-4c32-b52d-4f2933cd8471"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064557 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-log-ovn\") pod \"77b25d9f-8183-4c32-b52d-4f2933cd8471\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064581 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-scripts\") pod \"77b25d9f-8183-4c32-b52d-4f2933cd8471\" (UID: \"77b25d9f-8183-4c32-b52d-4f2933cd8471\") " Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064656 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "77b25d9f-8183-4c32-b52d-4f2933cd8471" (UID: "77b25d9f-8183-4c32-b52d-4f2933cd8471"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.064985 4979 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.065005 4979 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.065013 4979 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/77b25d9f-8183-4c32-b52d-4f2933cd8471-var-run\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.065329 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "77b25d9f-8183-4c32-b52d-4f2933cd8471" (UID: "77b25d9f-8183-4c32-b52d-4f2933cd8471"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.065594 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-scripts" (OuterVolumeSpecName: "scripts") pod "77b25d9f-8183-4c32-b52d-4f2933cd8471" (UID: "77b25d9f-8183-4c32-b52d-4f2933cd8471"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.069702 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77b25d9f-8183-4c32-b52d-4f2933cd8471-kube-api-access-f8fmr" (OuterVolumeSpecName: "kube-api-access-f8fmr") pod "77b25d9f-8183-4c32-b52d-4f2933cd8471" (UID: "77b25d9f-8183-4c32-b52d-4f2933cd8471"). InnerVolumeSpecName "kube-api-access-f8fmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.167376 4979 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.167456 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8fmr\" (UniqueName: \"kubernetes.io/projected/77b25d9f-8183-4c32-b52d-4f2933cd8471-kube-api-access-f8fmr\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.167473 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/77b25d9f-8183-4c32-b52d-4f2933cd8471-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.578778 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-cv4bf-config-cknxb" event={"ID":"77b25d9f-8183-4c32-b52d-4f2933cd8471","Type":"ContainerDied","Data":"db3ddc62c92913898afb5a210ad74285a94ad33a1681d44fc02ca4fe578971a5"} Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.578824 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db3ddc62c92913898afb5a210ad74285a94ad33a1681d44fc02ca4fe578971a5" Dec 04 13:21:19 crc kubenswrapper[4979]: I1204 13:21:19.578870 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-cv4bf-config-cknxb" Dec 04 13:21:20 crc kubenswrapper[4979]: I1204 13:21:20.063638 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-cv4bf-config-cknxb"] Dec 04 13:21:20 crc kubenswrapper[4979]: I1204 13:21:20.074515 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-cv4bf-config-cknxb"] Dec 04 13:21:20 crc kubenswrapper[4979]: I1204 13:21:20.199375 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:21:20 crc kubenswrapper[4979]: E1204 13:21:20.199973 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:21:20 crc kubenswrapper[4979]: I1204 13:21:20.210432 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77b25d9f-8183-4c32-b52d-4f2933cd8471" path="/var/lib/kubelet/pods/77b25d9f-8183-4c32-b52d-4f2933cd8471/volumes" Dec 04 13:21:20 crc kubenswrapper[4979]: I1204 13:21:20.286333 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-db-sync-pscsr"] Dec 04 13:21:20 crc kubenswrapper[4979]: I1204 13:21:20.593189 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-pscsr" event={"ID":"7cef70eb-2406-49ad-9fe3-83a6786082b2","Type":"ContainerStarted","Data":"a01363ae44add941f6585005cd399324d9c8c5f3cdb7ea4eec825eadbebfee56"} Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.603734 4979 generic.go:334] "Generic (PLEG): container finished" podID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerID="5d522e1988c1bca659fd5bd74fff73c7129e131fc9cf8f562dc05d6b8f9c374b" exitCode=0 Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.603917 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-pscsr" event={"ID":"7cef70eb-2406-49ad-9fe3-83a6786082b2","Type":"ContainerDied","Data":"5d522e1988c1bca659fd5bd74fff73c7129e131fc9cf8f562dc05d6b8f9c374b"} Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.607724 4979 generic.go:334] "Generic (PLEG): container finished" podID="06bddff5-a672-4ac8-810d-3332a26a8344" containerID="16a9c62335bfcd5c14617e10febb89109ba8c1f5269d781f97ab41ee40069b01" exitCode=0 Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.607797 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-x86s4" event={"ID":"06bddff5-a672-4ac8-810d-3332a26a8344","Type":"ContainerDied","Data":"16a9c62335bfcd5c14617e10febb89109ba8c1f5269d781f97ab41ee40069b01"} Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.614449 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-rsyslog-5bm9b" event={"ID":"1d361aab-dd58-4fc9-aa52-738ff495f640","Type":"ContainerStarted","Data":"e2f6ce8c77e8e49a4c8821d36924ce77269d6972eddcc9d33604288f619638f8"} Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.615268 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:21 crc kubenswrapper[4979]: I1204 13:21:21.664625 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-rsyslog-5bm9b" podStartSLOduration=2.377368782 podStartE2EDuration="14.664603923s" podCreationTimestamp="2025-12-04 13:21:07 +0000 UTC" firstStartedPulling="2025-12-04 13:21:08.113467783 +0000 UTC m=+5892.387763587" lastFinishedPulling="2025-12-04 13:21:20.400702924 +0000 UTC m=+5904.674998728" observedRunningTime="2025-12-04 13:21:21.663505514 +0000 UTC m=+5905.937801318" watchObservedRunningTime="2025-12-04 13:21:21.664603923 +0000 UTC m=+5905.938899727" Dec 04 13:21:22 crc kubenswrapper[4979]: I1204 13:21:22.626524 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-pscsr" event={"ID":"7cef70eb-2406-49ad-9fe3-83a6786082b2","Type":"ContainerStarted","Data":"7f0cfdbe507a575fd67f39aba66b7a5a2539f999829a35e10c7d82742cc3490a"} Dec 04 13:21:22 crc kubenswrapper[4979]: I1204 13:21:22.644503 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-db-sync-pscsr" podStartSLOduration=9.644482081 podStartE2EDuration="9.644482081s" podCreationTimestamp="2025-12-04 13:21:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:21:22.63991604 +0000 UTC m=+5906.914211874" watchObservedRunningTime="2025-12-04 13:21:22.644482081 +0000 UTC m=+5906.918777885" Dec 04 13:21:23 crc kubenswrapper[4979]: I1204 13:21:23.637943 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-x86s4" event={"ID":"06bddff5-a672-4ac8-810d-3332a26a8344","Type":"ContainerStarted","Data":"761b4d1abe60245e5b98dc10b689c438d24fce94b1a84dcdcce6a7ecb3d24cd5"} Dec 04 13:21:23 crc kubenswrapper[4979]: I1204 13:21:23.662470 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-x86s4" podStartSLOduration=2.871420077 podStartE2EDuration="16.662449072s" podCreationTimestamp="2025-12-04 13:21:07 +0000 UTC" firstStartedPulling="2025-12-04 13:21:08.711910353 +0000 UTC m=+5892.986206157" lastFinishedPulling="2025-12-04 13:21:22.502939348 +0000 UTC m=+5906.777235152" observedRunningTime="2025-12-04 13:21:23.653654009 +0000 UTC m=+5907.927949813" watchObservedRunningTime="2025-12-04 13:21:23.662449072 +0000 UTC m=+5907.936744876" Dec 04 13:21:27 crc kubenswrapper[4979]: I1204 13:21:27.674962 4979 generic.go:334] "Generic (PLEG): container finished" podID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerID="7f0cfdbe507a575fd67f39aba66b7a5a2539f999829a35e10c7d82742cc3490a" exitCode=0 Dec 04 13:21:27 crc kubenswrapper[4979]: I1204 13:21:27.675060 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-pscsr" event={"ID":"7cef70eb-2406-49ad-9fe3-83a6786082b2","Type":"ContainerDied","Data":"7f0cfdbe507a575fd67f39aba66b7a5a2539f999829a35e10c7d82742cc3490a"} Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.038249 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.174775 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data\") pod \"7cef70eb-2406-49ad-9fe3-83a6786082b2\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.174850 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-combined-ca-bundle\") pod \"7cef70eb-2406-49ad-9fe3-83a6786082b2\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.175016 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-scripts\") pod \"7cef70eb-2406-49ad-9fe3-83a6786082b2\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.175088 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data-merged\") pod \"7cef70eb-2406-49ad-9fe3-83a6786082b2\" (UID: \"7cef70eb-2406-49ad-9fe3-83a6786082b2\") " Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.180798 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data" (OuterVolumeSpecName: "config-data") pod "7cef70eb-2406-49ad-9fe3-83a6786082b2" (UID: "7cef70eb-2406-49ad-9fe3-83a6786082b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.180871 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-scripts" (OuterVolumeSpecName: "scripts") pod "7cef70eb-2406-49ad-9fe3-83a6786082b2" (UID: "7cef70eb-2406-49ad-9fe3-83a6786082b2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.204199 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "7cef70eb-2406-49ad-9fe3-83a6786082b2" (UID: "7cef70eb-2406-49ad-9fe3-83a6786082b2"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.206270 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7cef70eb-2406-49ad-9fe3-83a6786082b2" (UID: "7cef70eb-2406-49ad-9fe3-83a6786082b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.278855 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.278888 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.278899 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7cef70eb-2406-49ad-9fe3-83a6786082b2-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.278908 4979 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/7cef70eb-2406-49ad-9fe3-83a6786082b2-config-data-merged\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.697976 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-db-sync-pscsr" event={"ID":"7cef70eb-2406-49ad-9fe3-83a6786082b2","Type":"ContainerDied","Data":"a01363ae44add941f6585005cd399324d9c8c5f3cdb7ea4eec825eadbebfee56"} Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.698023 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a01363ae44add941f6585005cd399324d9c8c5f3cdb7ea4eec825eadbebfee56" Dec 04 13:21:29 crc kubenswrapper[4979]: I1204 13:21:29.698093 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-db-sync-pscsr" Dec 04 13:21:35 crc kubenswrapper[4979]: I1204 13:21:35.199459 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:21:35 crc kubenswrapper[4979]: E1204 13:21:35.200142 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:21:37 crc kubenswrapper[4979]: I1204 13:21:37.558629 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-rsyslog-5bm9b" Dec 04 13:21:49 crc kubenswrapper[4979]: I1204 13:21:49.199815 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:21:49 crc kubenswrapper[4979]: E1204 13:21:49.200775 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.526315 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-x86s4"] Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.527177 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/octavia-image-upload-59f8cff499-x86s4" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" containerName="octavia-amphora-httpd" containerID="cri-o://761b4d1abe60245e5b98dc10b689c438d24fce94b1a84dcdcce6a7ecb3d24cd5" gracePeriod=30 Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.952934 4979 generic.go:334] "Generic (PLEG): container finished" podID="06bddff5-a672-4ac8-810d-3332a26a8344" containerID="761b4d1abe60245e5b98dc10b689c438d24fce94b1a84dcdcce6a7ecb3d24cd5" exitCode=0 Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.953029 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-x86s4" event={"ID":"06bddff5-a672-4ac8-810d-3332a26a8344","Type":"ContainerDied","Data":"761b4d1abe60245e5b98dc10b689c438d24fce94b1a84dcdcce6a7ecb3d24cd5"} Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.953394 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-x86s4" event={"ID":"06bddff5-a672-4ac8-810d-3332a26a8344","Type":"ContainerDied","Data":"c906eb6793b80f3f6420b497a06e08985500bf6f6d6d4b21aeec6a408e0dcccf"} Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.953416 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c906eb6793b80f3f6420b497a06e08985500bf6f6d6d4b21aeec6a408e0dcccf" Dec 04 13:21:54 crc kubenswrapper[4979]: I1204 13:21:54.989438 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.071785 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/06bddff5-a672-4ac8-810d-3332a26a8344-httpd-config\") pod \"06bddff5-a672-4ac8-810d-3332a26a8344\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.072542 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/06bddff5-a672-4ac8-810d-3332a26a8344-amphora-image\") pod \"06bddff5-a672-4ac8-810d-3332a26a8344\" (UID: \"06bddff5-a672-4ac8-810d-3332a26a8344\") " Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.100505 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06bddff5-a672-4ac8-810d-3332a26a8344-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "06bddff5-a672-4ac8-810d-3332a26a8344" (UID: "06bddff5-a672-4ac8-810d-3332a26a8344"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.165120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06bddff5-a672-4ac8-810d-3332a26a8344-amphora-image" (OuterVolumeSpecName: "amphora-image") pod "06bddff5-a672-4ac8-810d-3332a26a8344" (UID: "06bddff5-a672-4ac8-810d-3332a26a8344"). InnerVolumeSpecName "amphora-image". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.175210 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/06bddff5-a672-4ac8-810d-3332a26a8344-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.175241 4979 reconciler_common.go:293] "Volume detached for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/06bddff5-a672-4ac8-810d-3332a26a8344-amphora-image\") on node \"crc\" DevicePath \"\"" Dec 04 13:21:55 crc kubenswrapper[4979]: I1204 13:21:55.964120 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-x86s4" Dec 04 13:21:56 crc kubenswrapper[4979]: I1204 13:21:56.001660 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-x86s4"] Dec 04 13:21:56 crc kubenswrapper[4979]: I1204 13:21:56.010992 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-x86s4"] Dec 04 13:21:56 crc kubenswrapper[4979]: I1204 13:21:56.210438 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" path="/var/lib/kubelet/pods/06bddff5-a672-4ac8-810d-3332a26a8344/volumes" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.464239 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-image-upload-59f8cff499-ghhlt"] Dec 04 13:21:58 crc kubenswrapper[4979]: E1204 13:21:58.466727 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerName="octavia-db-sync" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.466753 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerName="octavia-db-sync" Dec 04 13:21:58 crc kubenswrapper[4979]: E1204 13:21:58.466774 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" containerName="octavia-amphora-httpd" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.466781 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" containerName="octavia-amphora-httpd" Dec 04 13:21:58 crc kubenswrapper[4979]: E1204 13:21:58.466794 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerName="init" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.466803 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerName="init" Dec 04 13:21:58 crc kubenswrapper[4979]: E1204 13:21:58.466827 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77b25d9f-8183-4c32-b52d-4f2933cd8471" containerName="ovn-config" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.466834 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="77b25d9f-8183-4c32-b52d-4f2933cd8471" containerName="ovn-config" Dec 04 13:21:58 crc kubenswrapper[4979]: E1204 13:21:58.466870 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" containerName="init" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.466881 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" containerName="init" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.467089 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="06bddff5-a672-4ac8-810d-3332a26a8344" containerName="octavia-amphora-httpd" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.467111 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cef70eb-2406-49ad-9fe3-83a6786082b2" containerName="octavia-db-sync" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.467129 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="77b25d9f-8183-4c32-b52d-4f2933cd8471" containerName="ovn-config" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.468162 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.470924 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-config-data" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.480968 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-ghhlt"] Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.538038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/abea20e4-7ff1-4f96-abcd-8c0a86c7df7e-amphora-image\") pod \"octavia-image-upload-59f8cff499-ghhlt\" (UID: \"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e\") " pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.538233 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/abea20e4-7ff1-4f96-abcd-8c0a86c7df7e-httpd-config\") pod \"octavia-image-upload-59f8cff499-ghhlt\" (UID: \"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e\") " pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.639797 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/abea20e4-7ff1-4f96-abcd-8c0a86c7df7e-httpd-config\") pod \"octavia-image-upload-59f8cff499-ghhlt\" (UID: \"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e\") " pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.639956 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/abea20e4-7ff1-4f96-abcd-8c0a86c7df7e-amphora-image\") pod \"octavia-image-upload-59f8cff499-ghhlt\" (UID: \"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e\") " pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.640402 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-image\" (UniqueName: \"kubernetes.io/empty-dir/abea20e4-7ff1-4f96-abcd-8c0a86c7df7e-amphora-image\") pod \"octavia-image-upload-59f8cff499-ghhlt\" (UID: \"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e\") " pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.655668 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/abea20e4-7ff1-4f96-abcd-8c0a86c7df7e-httpd-config\") pod \"octavia-image-upload-59f8cff499-ghhlt\" (UID: \"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e\") " pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:58 crc kubenswrapper[4979]: I1204 13:21:58.789000 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" Dec 04 13:21:59 crc kubenswrapper[4979]: I1204 13:21:59.259676 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-image-upload-59f8cff499-ghhlt"] Dec 04 13:22:00 crc kubenswrapper[4979]: I1204 13:22:00.012235 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" event={"ID":"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e","Type":"ContainerStarted","Data":"d9984da331a2b3f0efad25e234db57db60b6bb2030cebbf45fbf869774cfeeb4"} Dec 04 13:22:00 crc kubenswrapper[4979]: I1204 13:22:00.013581 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" event={"ID":"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e","Type":"ContainerStarted","Data":"ad30acaf9b43b46575004d5bb71ca424e6183f9da4136ed5c6235a1acf1c652a"} Dec 04 13:22:00 crc kubenswrapper[4979]: I1204 13:22:00.199508 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:22:00 crc kubenswrapper[4979]: E1204 13:22:00.199806 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:22:01 crc kubenswrapper[4979]: I1204 13:22:01.023536 4979 generic.go:334] "Generic (PLEG): container finished" podID="abea20e4-7ff1-4f96-abcd-8c0a86c7df7e" containerID="d9984da331a2b3f0efad25e234db57db60b6bb2030cebbf45fbf869774cfeeb4" exitCode=0 Dec 04 13:22:01 crc kubenswrapper[4979]: I1204 13:22:01.023604 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" event={"ID":"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e","Type":"ContainerDied","Data":"d9984da331a2b3f0efad25e234db57db60b6bb2030cebbf45fbf869774cfeeb4"} Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.044326 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" event={"ID":"abea20e4-7ff1-4f96-abcd-8c0a86c7df7e","Type":"ContainerStarted","Data":"8e3e1dbd375ee6a0202bec5b86c1c384d83cb91671116df3adcdf8b4e5ac190f"} Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.067527 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-image-upload-59f8cff499-ghhlt" podStartSLOduration=2.018648464 podStartE2EDuration="5.067507884s" podCreationTimestamp="2025-12-04 13:21:58 +0000 UTC" firstStartedPulling="2025-12-04 13:21:59.263606712 +0000 UTC m=+5943.537902516" lastFinishedPulling="2025-12-04 13:22:02.312466132 +0000 UTC m=+5946.586761936" observedRunningTime="2025-12-04 13:22:03.058993248 +0000 UTC m=+5947.333289072" watchObservedRunningTime="2025-12-04 13:22:03.067507884 +0000 UTC m=+5947.341803688" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.758383 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sk98z"] Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.775157 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.804217 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk98z"] Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.853416 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8czg7\" (UniqueName: \"kubernetes.io/projected/475dc2d2-4009-4747-ae58-250ef35e725d-kube-api-access-8czg7\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.853472 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-utilities\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.853530 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-catalog-content\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.955369 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8czg7\" (UniqueName: \"kubernetes.io/projected/475dc2d2-4009-4747-ae58-250ef35e725d-kube-api-access-8czg7\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.955708 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-utilities\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.955787 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-catalog-content\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.956463 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-catalog-content\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.956584 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-utilities\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:03 crc kubenswrapper[4979]: I1204 13:22:03.976911 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8czg7\" (UniqueName: \"kubernetes.io/projected/475dc2d2-4009-4747-ae58-250ef35e725d-kube-api-access-8czg7\") pod \"redhat-marketplace-sk98z\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:04 crc kubenswrapper[4979]: I1204 13:22:04.117870 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:04 crc kubenswrapper[4979]: I1204 13:22:04.587266 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk98z"] Dec 04 13:22:04 crc kubenswrapper[4979]: W1204 13:22:04.588278 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod475dc2d2_4009_4747_ae58_250ef35e725d.slice/crio-1419afabd427ea09d81cf80ba4c5f836fa782316ef7caace4ce7dd47d11bc00f WatchSource:0}: Error finding container 1419afabd427ea09d81cf80ba4c5f836fa782316ef7caace4ce7dd47d11bc00f: Status 404 returned error can't find the container with id 1419afabd427ea09d81cf80ba4c5f836fa782316ef7caace4ce7dd47d11bc00f Dec 04 13:22:05 crc kubenswrapper[4979]: I1204 13:22:05.064664 4979 generic.go:334] "Generic (PLEG): container finished" podID="475dc2d2-4009-4747-ae58-250ef35e725d" containerID="0dd23969dd926382c1a538297fbf0aefa6cb8a53b5a890f0234b06f4531ada93" exitCode=0 Dec 04 13:22:05 crc kubenswrapper[4979]: I1204 13:22:05.064735 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk98z" event={"ID":"475dc2d2-4009-4747-ae58-250ef35e725d","Type":"ContainerDied","Data":"0dd23969dd926382c1a538297fbf0aefa6cb8a53b5a890f0234b06f4531ada93"} Dec 04 13:22:05 crc kubenswrapper[4979]: I1204 13:22:05.064958 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk98z" event={"ID":"475dc2d2-4009-4747-ae58-250ef35e725d","Type":"ContainerStarted","Data":"1419afabd427ea09d81cf80ba4c5f836fa782316ef7caace4ce7dd47d11bc00f"} Dec 04 13:22:07 crc kubenswrapper[4979]: I1204 13:22:07.093648 4979 generic.go:334] "Generic (PLEG): container finished" podID="475dc2d2-4009-4747-ae58-250ef35e725d" containerID="5f476ed15d807b99c40490adf0d3b0854035ef403de4238fc3a74b2cc472c7b7" exitCode=0 Dec 04 13:22:07 crc kubenswrapper[4979]: I1204 13:22:07.093736 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk98z" event={"ID":"475dc2d2-4009-4747-ae58-250ef35e725d","Type":"ContainerDied","Data":"5f476ed15d807b99c40490adf0d3b0854035ef403de4238fc3a74b2cc472c7b7"} Dec 04 13:22:08 crc kubenswrapper[4979]: I1204 13:22:08.105627 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk98z" event={"ID":"475dc2d2-4009-4747-ae58-250ef35e725d","Type":"ContainerStarted","Data":"1ddf66431762df53d8444c2674d249179393e4069e07d2771f4b302a9c9adc0a"} Dec 04 13:22:08 crc kubenswrapper[4979]: I1204 13:22:08.125044 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sk98z" podStartSLOduration=2.610450465 podStartE2EDuration="5.125020411s" podCreationTimestamp="2025-12-04 13:22:03 +0000 UTC" firstStartedPulling="2025-12-04 13:22:05.0666949 +0000 UTC m=+5949.340990724" lastFinishedPulling="2025-12-04 13:22:07.581264836 +0000 UTC m=+5951.855560670" observedRunningTime="2025-12-04 13:22:08.123238794 +0000 UTC m=+5952.397534598" watchObservedRunningTime="2025-12-04 13:22:08.125020411 +0000 UTC m=+5952.399316205" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.868518 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-healthmanager-ktcdr"] Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.870682 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.873610 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-certs-secret" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.873816 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-config-data" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.874093 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-healthmanager-scripts" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.880186 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-ktcdr"] Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.922860 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-amphora-certs\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.922926 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-config-data\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.923098 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-config-data-merged\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.923180 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-combined-ca-bundle\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.923230 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-scripts\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:11 crc kubenswrapper[4979]: I1204 13:22:11.923406 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-hm-ports\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.025163 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-hm-ports\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.025324 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-amphora-certs\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.025365 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-config-data\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.025407 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-config-data-merged\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.025443 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-combined-ca-bundle\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.025473 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-scripts\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.026394 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-config-data-merged\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.027076 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-hm-ports\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.032366 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-amphora-certs\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.032365 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-combined-ca-bundle\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.032601 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-scripts\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.047338 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf209239-4f9e-43dd-b5e0-fc1253c4fd7a-config-data\") pod \"octavia-healthmanager-ktcdr\" (UID: \"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a\") " pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:12 crc kubenswrapper[4979]: I1204 13:22:12.190670 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.211057 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-housekeeping-fmc22"] Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.213219 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.215126 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-config-data" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.217192 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-housekeeping-scripts" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.224894 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-fmc22"] Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.358555 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-amphora-certs\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.358845 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f724c253-8e48-4d2c-ae4a-f24351b450a0-config-data-merged\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.359048 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-combined-ca-bundle\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.359087 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-config-data\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.359116 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-scripts\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.359226 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f724c253-8e48-4d2c-ae4a-f24351b450a0-hm-ports\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.460535 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f724c253-8e48-4d2c-ae4a-f24351b450a0-hm-ports\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.460636 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-amphora-certs\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.460699 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f724c253-8e48-4d2c-ae4a-f24351b450a0-config-data-merged\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.460752 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-combined-ca-bundle\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.460770 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-config-data\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.460789 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-scripts\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.461668 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f724c253-8e48-4d2c-ae4a-f24351b450a0-config-data-merged\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.462030 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/f724c253-8e48-4d2c-ae4a-f24351b450a0-hm-ports\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.480874 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-amphora-certs\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.480942 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-combined-ca-bundle\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.481034 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-scripts\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.482672 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f724c253-8e48-4d2c-ae4a-f24351b450a0-config-data\") pod \"octavia-housekeeping-fmc22\" (UID: \"f724c253-8e48-4d2c-ae4a-f24351b450a0\") " pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.533049 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:13 crc kubenswrapper[4979]: I1204 13:22:13.701002 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-healthmanager-ktcdr"] Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.122925 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.123270 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.138549 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-housekeeping-fmc22"] Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.162086 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-ktcdr" event={"ID":"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a","Type":"ContainerStarted","Data":"29e24447987e911eae936ebf81bf1981678fdc58d3f9c2a0dafeac20768ffda9"} Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.198688 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.255501 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.349426 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/octavia-worker-w2lng"] Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.351199 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.353341 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-config-data" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.354266 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"octavia-worker-scripts" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.360671 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-w2lng"] Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.440945 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk98z"] Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.480048 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6efdadb0-0d97-4a27-9fa4-df4610542d61-hm-ports\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.480251 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-amphora-certs\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.480286 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-config-data\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.480745 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-combined-ca-bundle\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.480849 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-scripts\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.481161 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6efdadb0-0d97-4a27-9fa4-df4610542d61-config-data-merged\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.582818 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-scripts\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.583887 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6efdadb0-0d97-4a27-9fa4-df4610542d61-config-data-merged\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.583970 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6efdadb0-0d97-4a27-9fa4-df4610542d61-hm-ports\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.584083 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-amphora-certs\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.584120 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-config-data\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.584318 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-combined-ca-bundle\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.585969 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hm-ports\" (UniqueName: \"kubernetes.io/configmap/6efdadb0-0d97-4a27-9fa4-df4610542d61-hm-ports\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.593064 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-scripts\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.593285 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-combined-ca-bundle\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.593366 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"amphora-certs\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-amphora-certs\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.692424 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/6efdadb0-0d97-4a27-9fa4-df4610542d61-config-data-merged\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.695217 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6efdadb0-0d97-4a27-9fa4-df4610542d61-config-data\") pod \"octavia-worker-w2lng\" (UID: \"6efdadb0-0d97-4a27-9fa4-df4610542d61\") " pod="openstack/octavia-worker-w2lng" Dec 04 13:22:14 crc kubenswrapper[4979]: I1204 13:22:14.974570 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/octavia-worker-w2lng" Dec 04 13:22:15 crc kubenswrapper[4979]: I1204 13:22:15.179139 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-fmc22" event={"ID":"f724c253-8e48-4d2c-ae4a-f24351b450a0","Type":"ContainerStarted","Data":"776871f5f0e913ec78e9f547b2a0fb13de3a4d5285017d2fab8fc0ee7a789100"} Dec 04 13:22:15 crc kubenswrapper[4979]: I1204 13:22:15.180892 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-ktcdr" event={"ID":"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a","Type":"ContainerStarted","Data":"25f499042f119466f5365cd8d276a855c055d8e95f2b3e20bcc51631e8368864"} Dec 04 13:22:15 crc kubenswrapper[4979]: I1204 13:22:15.198956 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:22:15 crc kubenswrapper[4979]: E1204 13:22:15.199221 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:22:15 crc kubenswrapper[4979]: I1204 13:22:15.502580 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/octavia-worker-w2lng"] Dec 04 13:22:16 crc kubenswrapper[4979]: I1204 13:22:16.192683 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w2lng" event={"ID":"6efdadb0-0d97-4a27-9fa4-df4610542d61","Type":"ContainerStarted","Data":"f22e506f9a58d0c1094b4ae1bbd1e672bcc8ce35b10e239e41e7eb7e10b1325b"} Dec 04 13:22:16 crc kubenswrapper[4979]: I1204 13:22:16.192917 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sk98z" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="registry-server" containerID="cri-o://1ddf66431762df53d8444c2674d249179393e4069e07d2771f4b302a9c9adc0a" gracePeriod=2 Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.204204 4979 generic.go:334] "Generic (PLEG): container finished" podID="475dc2d2-4009-4747-ae58-250ef35e725d" containerID="1ddf66431762df53d8444c2674d249179393e4069e07d2771f4b302a9c9adc0a" exitCode=0 Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.204310 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk98z" event={"ID":"475dc2d2-4009-4747-ae58-250ef35e725d","Type":"ContainerDied","Data":"1ddf66431762df53d8444c2674d249179393e4069e07d2771f4b302a9c9adc0a"} Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.423673 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.547545 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8czg7\" (UniqueName: \"kubernetes.io/projected/475dc2d2-4009-4747-ae58-250ef35e725d-kube-api-access-8czg7\") pod \"475dc2d2-4009-4747-ae58-250ef35e725d\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.547603 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-utilities\") pod \"475dc2d2-4009-4747-ae58-250ef35e725d\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.548025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-catalog-content\") pod \"475dc2d2-4009-4747-ae58-250ef35e725d\" (UID: \"475dc2d2-4009-4747-ae58-250ef35e725d\") " Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.548716 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-utilities" (OuterVolumeSpecName: "utilities") pod "475dc2d2-4009-4747-ae58-250ef35e725d" (UID: "475dc2d2-4009-4747-ae58-250ef35e725d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.556034 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/475dc2d2-4009-4747-ae58-250ef35e725d-kube-api-access-8czg7" (OuterVolumeSpecName: "kube-api-access-8czg7") pod "475dc2d2-4009-4747-ae58-250ef35e725d" (UID: "475dc2d2-4009-4747-ae58-250ef35e725d"). InnerVolumeSpecName "kube-api-access-8czg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.568166 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "475dc2d2-4009-4747-ae58-250ef35e725d" (UID: "475dc2d2-4009-4747-ae58-250ef35e725d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.650411 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.650443 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8czg7\" (UniqueName: \"kubernetes.io/projected/475dc2d2-4009-4747-ae58-250ef35e725d-kube-api-access-8czg7\") on node \"crc\" DevicePath \"\"" Dec 04 13:22:17 crc kubenswrapper[4979]: I1204 13:22:17.650454 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/475dc2d2-4009-4747-ae58-250ef35e725d-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.213364 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w2lng" event={"ID":"6efdadb0-0d97-4a27-9fa4-df4610542d61","Type":"ContainerStarted","Data":"58586b2789e9d8d7a3cff00ea7bf88541126685263212153d41783151d6efbc6"} Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.215591 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-fmc22" event={"ID":"f724c253-8e48-4d2c-ae4a-f24351b450a0","Type":"ContainerStarted","Data":"fdbdbaf9f4934e8246e09e4e5de4f6c2b58ca2a3a9e2787f52fa9316cd0d8cbf"} Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.220735 4979 generic.go:334] "Generic (PLEG): container finished" podID="cf209239-4f9e-43dd-b5e0-fc1253c4fd7a" containerID="25f499042f119466f5365cd8d276a855c055d8e95f2b3e20bcc51631e8368864" exitCode=0 Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.220913 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-ktcdr" event={"ID":"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a","Type":"ContainerDied","Data":"25f499042f119466f5365cd8d276a855c055d8e95f2b3e20bcc51631e8368864"} Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.226100 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sk98z" event={"ID":"475dc2d2-4009-4747-ae58-250ef35e725d","Type":"ContainerDied","Data":"1419afabd427ea09d81cf80ba4c5f836fa782316ef7caace4ce7dd47d11bc00f"} Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.226166 4979 scope.go:117] "RemoveContainer" containerID="1ddf66431762df53d8444c2674d249179393e4069e07d2771f4b302a9c9adc0a" Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.226415 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sk98z" Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.346069 4979 scope.go:117] "RemoveContainer" containerID="5f476ed15d807b99c40490adf0d3b0854035ef403de4238fc3a74b2cc472c7b7" Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.399883 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk98z"] Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.416982 4979 scope.go:117] "RemoveContainer" containerID="0dd23969dd926382c1a538297fbf0aefa6cb8a53b5a890f0234b06f4531ada93" Dec 04 13:22:18 crc kubenswrapper[4979]: I1204 13:22:18.425822 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sk98z"] Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.238874 4979 generic.go:334] "Generic (PLEG): container finished" podID="6efdadb0-0d97-4a27-9fa4-df4610542d61" containerID="58586b2789e9d8d7a3cff00ea7bf88541126685263212153d41783151d6efbc6" exitCode=0 Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.239563 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w2lng" event={"ID":"6efdadb0-0d97-4a27-9fa4-df4610542d61","Type":"ContainerDied","Data":"58586b2789e9d8d7a3cff00ea7bf88541126685263212153d41783151d6efbc6"} Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.246288 4979 generic.go:334] "Generic (PLEG): container finished" podID="f724c253-8e48-4d2c-ae4a-f24351b450a0" containerID="fdbdbaf9f4934e8246e09e4e5de4f6c2b58ca2a3a9e2787f52fa9316cd0d8cbf" exitCode=0 Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.246351 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-fmc22" event={"ID":"f724c253-8e48-4d2c-ae4a-f24351b450a0","Type":"ContainerDied","Data":"fdbdbaf9f4934e8246e09e4e5de4f6c2b58ca2a3a9e2787f52fa9316cd0d8cbf"} Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.252055 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-healthmanager-ktcdr" event={"ID":"cf209239-4f9e-43dd-b5e0-fc1253c4fd7a","Type":"ContainerStarted","Data":"b2c8144bd223229071fefd3b2804b03cf4b3ed511a67866183c8c995d26467b8"} Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.252467 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:19 crc kubenswrapper[4979]: I1204 13:22:19.322653 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-healthmanager-ktcdr" podStartSLOduration=8.322630365 podStartE2EDuration="8.322630365s" podCreationTimestamp="2025-12-04 13:22:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:22:19.286567186 +0000 UTC m=+5963.560862990" watchObservedRunningTime="2025-12-04 13:22:19.322630365 +0000 UTC m=+5963.596926169" Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.211932 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" path="/var/lib/kubelet/pods/475dc2d2-4009-4747-ae58-250ef35e725d/volumes" Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.264970 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-housekeeping-fmc22" event={"ID":"f724c253-8e48-4d2c-ae4a-f24351b450a0","Type":"ContainerStarted","Data":"2fbd35ba5bdeb0fda3e7a1fb4d25c90dd87233cde06c9a69fb10824da568f476"} Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.265070 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.270058 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/octavia-worker-w2lng" event={"ID":"6efdadb0-0d97-4a27-9fa4-df4610542d61","Type":"ContainerStarted","Data":"9bfbc948b3ba02366def914ea22dfdc4e42ec1622a5eea6d0db913c86bc0495a"} Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.270101 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/octavia-worker-w2lng" Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.290099 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-housekeeping-fmc22" podStartSLOduration=4.193089124 podStartE2EDuration="7.290074044s" podCreationTimestamp="2025-12-04 13:22:13 +0000 UTC" firstStartedPulling="2025-12-04 13:22:14.172545536 +0000 UTC m=+5958.446841340" lastFinishedPulling="2025-12-04 13:22:17.269530456 +0000 UTC m=+5961.543826260" observedRunningTime="2025-12-04 13:22:20.280946691 +0000 UTC m=+5964.555242495" watchObservedRunningTime="2025-12-04 13:22:20.290074044 +0000 UTC m=+5964.564369868" Dec 04 13:22:20 crc kubenswrapper[4979]: I1204 13:22:20.316027 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/octavia-worker-w2lng" podStartSLOduration=4.557477995 podStartE2EDuration="6.316010653s" podCreationTimestamp="2025-12-04 13:22:14 +0000 UTC" firstStartedPulling="2025-12-04 13:22:15.514288676 +0000 UTC m=+5959.788584480" lastFinishedPulling="2025-12-04 13:22:17.272821334 +0000 UTC m=+5961.547117138" observedRunningTime="2025-12-04 13:22:20.307767713 +0000 UTC m=+5964.582063517" watchObservedRunningTime="2025-12-04 13:22:20.316010653 +0000 UTC m=+5964.590306457" Dec 04 13:22:26 crc kubenswrapper[4979]: I1204 13:22:26.206186 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:22:26 crc kubenswrapper[4979]: E1204 13:22:26.207961 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:22:27 crc kubenswrapper[4979]: I1204 13:22:27.219691 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-healthmanager-ktcdr" Dec 04 13:22:28 crc kubenswrapper[4979]: I1204 13:22:28.564100 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-housekeeping-fmc22" Dec 04 13:22:30 crc kubenswrapper[4979]: I1204 13:22:30.005947 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/octavia-worker-w2lng" Dec 04 13:22:41 crc kubenswrapper[4979]: I1204 13:22:41.199054 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:22:41 crc kubenswrapper[4979]: I1204 13:22:41.460718 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"be34e428814c510aac7c63fa5d7dc77256bd5b17bcd712eed8e9b57775e43b6a"} Dec 04 13:22:48 crc kubenswrapper[4979]: I1204 13:22:48.232068 4979 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod475dc2d2-4009-4747-ae58-250ef35e725d"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod475dc2d2-4009-4747-ae58-250ef35e725d] : Timed out while waiting for systemd to remove kubepods-burstable-pod475dc2d2_4009_4747_ae58_250ef35e725d.slice" Dec 04 13:23:00 crc kubenswrapper[4979]: I1204 13:23:00.055794 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-bf3c-account-create-update-k78d4"] Dec 04 13:23:00 crc kubenswrapper[4979]: I1204 13:23:00.069797 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-8snsg"] Dec 04 13:23:00 crc kubenswrapper[4979]: I1204 13:23:00.082815 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-bf3c-account-create-update-k78d4"] Dec 04 13:23:00 crc kubenswrapper[4979]: I1204 13:23:00.095896 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-8snsg"] Dec 04 13:23:00 crc kubenswrapper[4979]: I1204 13:23:00.232798 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35d27ea1-b55f-4072-af5d-7dfa136317a2" path="/var/lib/kubelet/pods/35d27ea1-b55f-4072-af5d-7dfa136317a2/volumes" Dec 04 13:23:00 crc kubenswrapper[4979]: I1204 13:23:00.240549 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4bd880a-4a20-4038-a98a-7fc2c170c739" path="/var/lib/kubelet/pods/a4bd880a-4a20-4038-a98a-7fc2c170c739/volumes" Dec 04 13:23:06 crc kubenswrapper[4979]: I1204 13:23:06.037041 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-srlgj"] Dec 04 13:23:06 crc kubenswrapper[4979]: I1204 13:23:06.050425 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-srlgj"] Dec 04 13:23:06 crc kubenswrapper[4979]: I1204 13:23:06.214460 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39e0503d-f5da-4fe8-8488-13e9e2834b8f" path="/var/lib/kubelet/pods/39e0503d-f5da-4fe8-8488-13e9e2834b8f/volumes" Dec 04 13:23:07 crc kubenswrapper[4979]: I1204 13:23:07.969103 4979 scope.go:117] "RemoveContainer" containerID="d6e01d06690e670f9a5f9acfd54c418b55bf49e11f41246df5d8a0706f012ccf" Dec 04 13:23:08 crc kubenswrapper[4979]: I1204 13:23:08.002826 4979 scope.go:117] "RemoveContainer" containerID="1da2e815fda862ad0ddafcd951202e197e5c13bd9f123d32d3ec9bc4c29e8edd" Dec 04 13:23:08 crc kubenswrapper[4979]: I1204 13:23:08.036350 4979 scope.go:117] "RemoveContainer" containerID="802a051e7514b076257335eed5d687cdda129a9f5c03f81043742d58e215b285" Dec 04 13:23:08 crc kubenswrapper[4979]: I1204 13:23:08.062658 4979 scope.go:117] "RemoveContainer" containerID="4c9a940c9aefdc19229ce43bd16b871d1dbc397f9bbd271044a98121efd9b57c" Dec 04 13:23:08 crc kubenswrapper[4979]: I1204 13:23:08.098915 4979 scope.go:117] "RemoveContainer" containerID="dc559ab693d8c1733e9c4b638a86fe0dea3a042d875ae00c7203ba142adec4c0" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.736037 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-554cf9495c-tbhnf"] Dec 04 13:23:13 crc kubenswrapper[4979]: E1204 13:23:13.737735 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="extract-utilities" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.737765 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="extract-utilities" Dec 04 13:23:13 crc kubenswrapper[4979]: E1204 13:23:13.737798 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="extract-content" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.737808 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="extract-content" Dec 04 13:23:13 crc kubenswrapper[4979]: E1204 13:23:13.737856 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="registry-server" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.737865 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="registry-server" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.738125 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="475dc2d2-4009-4747-ae58-250ef35e725d" containerName="registry-server" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.739986 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.745109 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-pf9qk" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.745321 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.745388 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.745528 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.756992 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-554cf9495c-tbhnf"] Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.804490 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.804939 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-log" containerID="cri-o://c247c615c9dfd23d374587ba506083ce2c138d5972ecdca3fbcc960f05b4f49b" gracePeriod=30 Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.814284 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-httpd" containerID="cri-o://73d277822e69dc649a1304a039dab984e18070400cece29139dde97ae693b2d8" gracePeriod=30 Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.848728 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b774f74bf-v9dzj"] Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.851114 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.875062 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-config-data\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.875706 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-horizon-secret-key\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.875799 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-logs\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.876038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-scripts\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.876072 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dk2q\" (UniqueName: \"kubernetes.io/projected/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-kube-api-access-5dk2q\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.912600 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b774f74bf-v9dzj"] Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.931389 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.931968 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-log" containerID="cri-o://2397952bdf020e37361aff8cdeedb643db349cada1efde5eb5f2216a061cf75d" gracePeriod=30 Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.932192 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-httpd" containerID="cri-o://f801e69ff572d809c6a02b485a314f214f9bcbfcbc2a040df5c53900140cae16" gracePeriod=30 Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978261 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-config-data\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978316 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-horizon-secret-key\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978355 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-logs\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978384 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/58966164-dd31-4e68-942e-97f6e9b78574-horizon-secret-key\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978421 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k72p\" (UniqueName: \"kubernetes.io/projected/58966164-dd31-4e68-942e-97f6e9b78574-kube-api-access-5k72p\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978449 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-scripts\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978474 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58966164-dd31-4e68-942e-97f6e9b78574-logs\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978528 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-config-data\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978546 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-scripts\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.978560 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dk2q\" (UniqueName: \"kubernetes.io/projected/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-kube-api-access-5dk2q\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.979908 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-config-data\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.981470 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-logs\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:13 crc kubenswrapper[4979]: I1204 13:23:13.981643 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-scripts\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:13.999998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-horizon-secret-key\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.000679 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dk2q\" (UniqueName: \"kubernetes.io/projected/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-kube-api-access-5dk2q\") pod \"horizon-554cf9495c-tbhnf\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.071539 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.079915 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k72p\" (UniqueName: \"kubernetes.io/projected/58966164-dd31-4e68-942e-97f6e9b78574-kube-api-access-5k72p\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.080000 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-scripts\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.080041 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58966164-dd31-4e68-942e-97f6e9b78574-logs\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.080129 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-config-data\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.080244 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/58966164-dd31-4e68-942e-97f6e9b78574-horizon-secret-key\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.081433 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58966164-dd31-4e68-942e-97f6e9b78574-logs\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.081656 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-scripts\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.082254 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-config-data\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.083894 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/58966164-dd31-4e68-942e-97f6e9b78574-horizon-secret-key\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.100348 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k72p\" (UniqueName: \"kubernetes.io/projected/58966164-dd31-4e68-942e-97f6e9b78574-kube-api-access-5k72p\") pod \"horizon-7b774f74bf-v9dzj\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.214404 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.423620 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-554cf9495c-tbhnf"] Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.465380 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-64f76457df-phmmr"] Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.470913 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.482392 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f76457df-phmmr"] Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.552128 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-554cf9495c-tbhnf"] Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.558568 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.589313 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/016a470f-300c-46b0-9241-f73f843de724-horizon-secret-key\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.589391 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-config-data\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.589537 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016a470f-300c-46b0-9241-f73f843de724-logs\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.589611 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6rm6\" (UniqueName: \"kubernetes.io/projected/016a470f-300c-46b0-9241-f73f843de724-kube-api-access-g6rm6\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.589710 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-scripts\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.691662 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/016a470f-300c-46b0-9241-f73f843de724-horizon-secret-key\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.691715 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-config-data\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.691761 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016a470f-300c-46b0-9241-f73f843de724-logs\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.691806 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6rm6\" (UniqueName: \"kubernetes.io/projected/016a470f-300c-46b0-9241-f73f843de724-kube-api-access-g6rm6\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.691902 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-scripts\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.692211 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016a470f-300c-46b0-9241-f73f843de724-logs\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.692779 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-scripts\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.693481 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-config-data\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.699364 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/016a470f-300c-46b0-9241-f73f843de724-horizon-secret-key\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.711277 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6rm6\" (UniqueName: \"kubernetes.io/projected/016a470f-300c-46b0-9241-f73f843de724-kube-api-access-g6rm6\") pod \"horizon-64f76457df-phmmr\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.726266 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b774f74bf-v9dzj"] Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.800544 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.815508 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b774f74bf-v9dzj" event={"ID":"58966164-dd31-4e68-942e-97f6e9b78574","Type":"ContainerStarted","Data":"9ec5ba78b907fd11744236267b125cdbe4a624537a901c0e6e14f497198a7274"} Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.819446 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6817c34-ecd6-4e01-8968-ad815733d564" containerID="2397952bdf020e37361aff8cdeedb643db349cada1efde5eb5f2216a061cf75d" exitCode=143 Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.819541 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6817c34-ecd6-4e01-8968-ad815733d564","Type":"ContainerDied","Data":"2397952bdf020e37361aff8cdeedb643db349cada1efde5eb5f2216a061cf75d"} Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.820986 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-554cf9495c-tbhnf" event={"ID":"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4","Type":"ContainerStarted","Data":"8966fcdabe19ce318304bbc3885c12a2579c7092608fda126190accbee60b561"} Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.823098 4979 generic.go:334] "Generic (PLEG): container finished" podID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerID="c247c615c9dfd23d374587ba506083ce2c138d5972ecdca3fbcc960f05b4f49b" exitCode=143 Dec 04 13:23:14 crc kubenswrapper[4979]: I1204 13:23:14.823124 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"513a2a30-d753-457f-b030-3a40a2a2f8eb","Type":"ContainerDied","Data":"c247c615c9dfd23d374587ba506083ce2c138d5972ecdca3fbcc960f05b4f49b"} Dec 04 13:23:15 crc kubenswrapper[4979]: I1204 13:23:15.262886 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-64f76457df-phmmr"] Dec 04 13:23:15 crc kubenswrapper[4979]: I1204 13:23:15.836506 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f76457df-phmmr" event={"ID":"016a470f-300c-46b0-9241-f73f843de724","Type":"ContainerStarted","Data":"132f429ae86e6a407aaedba6d1e26a0d92736155fa2157b0859c399ae23ff61e"} Dec 04 13:23:17 crc kubenswrapper[4979]: I1204 13:23:17.860333 4979 generic.go:334] "Generic (PLEG): container finished" podID="a6817c34-ecd6-4e01-8968-ad815733d564" containerID="f801e69ff572d809c6a02b485a314f214f9bcbfcbc2a040df5c53900140cae16" exitCode=0 Dec 04 13:23:17 crc kubenswrapper[4979]: I1204 13:23:17.860400 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6817c34-ecd6-4e01-8968-ad815733d564","Type":"ContainerDied","Data":"f801e69ff572d809c6a02b485a314f214f9bcbfcbc2a040df5c53900140cae16"} Dec 04 13:23:17 crc kubenswrapper[4979]: I1204 13:23:17.864944 4979 generic.go:334] "Generic (PLEG): container finished" podID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerID="73d277822e69dc649a1304a039dab984e18070400cece29139dde97ae693b2d8" exitCode=0 Dec 04 13:23:17 crc kubenswrapper[4979]: I1204 13:23:17.864992 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"513a2a30-d753-457f-b030-3a40a2a2f8eb","Type":"ContainerDied","Data":"73d277822e69dc649a1304a039dab984e18070400cece29139dde97ae693b2d8"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.484179 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.536852 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-combined-ca-bundle\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.536971 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-ceph\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.537042 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vspjb\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-kube-api-access-vspjb\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.537078 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-config-data\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.537114 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-scripts\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.544124 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-logs" (OuterVolumeSpecName: "logs") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.544194 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-logs\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.544537 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-httpd-run\") pod \"513a2a30-d753-457f-b030-3a40a2a2f8eb\" (UID: \"513a2a30-d753-457f-b030-3a40a2a2f8eb\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.545765 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.545809 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.546243 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-kube-api-access-vspjb" (OuterVolumeSpecName: "kube-api-access-vspjb") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "kube-api-access-vspjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.548768 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-scripts" (OuterVolumeSpecName: "scripts") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.566650 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-ceph" (OuterVolumeSpecName: "ceph") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.648516 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.649050 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vspjb\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-kube-api-access-vspjb\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.649070 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.649083 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/513a2a30-d753-457f-b030-3a40a2a2f8eb-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.649095 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.649107 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/513a2a30-d753-457f-b030-3a40a2a2f8eb-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.658713 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.671461 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-config-data" (OuterVolumeSpecName: "config-data") pod "513a2a30-d753-457f-b030-3a40a2a2f8eb" (UID: "513a2a30-d753-457f-b030-3a40a2a2f8eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751013 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-logs\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751097 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-scripts\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751146 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-ceph\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751249 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-httpd-run\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751377 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-config-data\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751436 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-combined-ca-bundle\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751481 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkgts\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-kube-api-access-lkgts\") pod \"a6817c34-ecd6-4e01-8968-ad815733d564\" (UID: \"a6817c34-ecd6-4e01-8968-ad815733d564\") " Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751706 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.751989 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/513a2a30-d753-457f-b030-3a40a2a2f8eb-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.752004 4979 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.752231 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-logs" (OuterVolumeSpecName: "logs") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.754991 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-ceph" (OuterVolumeSpecName: "ceph") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.758742 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-kube-api-access-lkgts" (OuterVolumeSpecName: "kube-api-access-lkgts") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "kube-api-access-lkgts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.759365 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-scripts" (OuterVolumeSpecName: "scripts") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.819507 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.853627 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkgts\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-kube-api-access-lkgts\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.853660 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a6817c34-ecd6-4e01-8968-ad815733d564-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.853670 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.853677 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/a6817c34-ecd6-4e01-8968-ad815733d564-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.853688 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.862537 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-config-data" (OuterVolumeSpecName: "config-data") pod "a6817c34-ecd6-4e01-8968-ad815733d564" (UID: "a6817c34-ecd6-4e01-8968-ad815733d564"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.927833 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f76457df-phmmr" event={"ID":"016a470f-300c-46b0-9241-f73f843de724","Type":"ContainerStarted","Data":"f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.927891 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f76457df-phmmr" event={"ID":"016a470f-300c-46b0-9241-f73f843de724","Type":"ContainerStarted","Data":"19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.930552 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.930567 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a6817c34-ecd6-4e01-8968-ad815733d564","Type":"ContainerDied","Data":"401e783b59f75fe92a9e431f4aea337f82d9a83e23dc30cf8e4b39ac26109cf2"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.930708 4979 scope.go:117] "RemoveContainer" containerID="f801e69ff572d809c6a02b485a314f214f9bcbfcbc2a040df5c53900140cae16" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.933938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-554cf9495c-tbhnf" event={"ID":"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4","Type":"ContainerStarted","Data":"98a9830f316748de1b45673ea3faad97c57602e0787223babafd4e8d47228034"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.933979 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-554cf9495c-tbhnf" event={"ID":"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4","Type":"ContainerStarted","Data":"22e6dbd3a9145842ea2090c72f861e332e0096c39dc9f53687a22e2321681a28"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.934095 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-554cf9495c-tbhnf" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon-log" containerID="cri-o://22e6dbd3a9145842ea2090c72f861e332e0096c39dc9f53687a22e2321681a28" gracePeriod=30 Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.934352 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-554cf9495c-tbhnf" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon" containerID="cri-o://98a9830f316748de1b45673ea3faad97c57602e0787223babafd4e8d47228034" gracePeriod=30 Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.937818 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"513a2a30-d753-457f-b030-3a40a2a2f8eb","Type":"ContainerDied","Data":"64b3d98276797fdbb752831b12c050c7fc0d8bb7373d0056739e7ec29fabace2"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.937911 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.956864 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a6817c34-ecd6-4e01-8968-ad815733d564-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.962672 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b774f74bf-v9dzj" event={"ID":"58966164-dd31-4e68-942e-97f6e9b78574","Type":"ContainerStarted","Data":"811bf09226a25589eb9a35ebd7ffb031a18426a23b7b85e6c854ef2e40fd0fed"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.962718 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b774f74bf-v9dzj" event={"ID":"58966164-dd31-4e68-942e-97f6e9b78574","Type":"ContainerStarted","Data":"f830632c42ca4df4c807861d9a6e71f1057a520dda6a44e249bf4828d2a5bbb1"} Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.963105 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-64f76457df-phmmr" podStartSLOduration=2.082673556 podStartE2EDuration="7.963081949s" podCreationTimestamp="2025-12-04 13:23:14 +0000 UTC" firstStartedPulling="2025-12-04 13:23:15.276221867 +0000 UTC m=+6019.550517671" lastFinishedPulling="2025-12-04 13:23:21.15663026 +0000 UTC m=+6025.430926064" observedRunningTime="2025-12-04 13:23:21.949436947 +0000 UTC m=+6026.223732751" watchObservedRunningTime="2025-12-04 13:23:21.963081949 +0000 UTC m=+6026.237377753" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.970286 4979 scope.go:117] "RemoveContainer" containerID="2397952bdf020e37361aff8cdeedb643db349cada1efde5eb5f2216a061cf75d" Dec 04 13:23:21 crc kubenswrapper[4979]: I1204 13:23:21.977556 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-554cf9495c-tbhnf" podStartSLOduration=2.402754273 podStartE2EDuration="8.977538714s" podCreationTimestamp="2025-12-04 13:23:13 +0000 UTC" firstStartedPulling="2025-12-04 13:23:14.558260552 +0000 UTC m=+6018.832556356" lastFinishedPulling="2025-12-04 13:23:21.133044993 +0000 UTC m=+6025.407340797" observedRunningTime="2025-12-04 13:23:21.974214495 +0000 UTC m=+6026.248510309" watchObservedRunningTime="2025-12-04 13:23:21.977538714 +0000 UTC m=+6026.251834518" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.003160 4979 scope.go:117] "RemoveContainer" containerID="73d277822e69dc649a1304a039dab984e18070400cece29139dde97ae693b2d8" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.010013 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.032331 4979 scope.go:117] "RemoveContainer" containerID="c247c615c9dfd23d374587ba506083ce2c138d5972ecdca3fbcc960f05b4f49b" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.043411 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.065248 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.077661 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088067 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: E1204 13:23:22.088540 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-log" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088559 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-log" Dec 04 13:23:22 crc kubenswrapper[4979]: E1204 13:23:22.088591 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-log" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088599 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-log" Dec 04 13:23:22 crc kubenswrapper[4979]: E1204 13:23:22.088635 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-httpd" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088644 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-httpd" Dec 04 13:23:22 crc kubenswrapper[4979]: E1204 13:23:22.088661 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-httpd" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088668 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-httpd" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088861 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-log" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088883 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" containerName="glance-httpd" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088894 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-log" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.088904 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" containerName="glance-httpd" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.090099 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.099028 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.099318 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gbs9f" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.099482 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.102724 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.105379 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.107713 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.149436 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b774f74bf-v9dzj" podStartSLOduration=2.717647013 podStartE2EDuration="9.149276709s" podCreationTimestamp="2025-12-04 13:23:13 +0000 UTC" firstStartedPulling="2025-12-04 13:23:14.723892335 +0000 UTC m=+6018.998188139" lastFinishedPulling="2025-12-04 13:23:21.155522031 +0000 UTC m=+6025.429817835" observedRunningTime="2025-12-04 13:23:22.053337459 +0000 UTC m=+6026.327633263" watchObservedRunningTime="2025-12-04 13:23:22.149276709 +0000 UTC m=+6026.423572513" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164496 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/838f0563-3ad1-4f68-9727-40140a0ba82a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164544 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164598 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-scripts\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164634 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-ceph\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164657 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164685 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164728 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-logs\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164758 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838f0563-3ad1-4f68-9727-40140a0ba82a-logs\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164786 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/838f0563-3ad1-4f68-9727-40140a0ba82a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164802 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-config-data\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164831 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164853 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx2gn\" (UniqueName: \"kubernetes.io/projected/838f0563-3ad1-4f68-9727-40140a0ba82a-kube-api-access-hx2gn\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164911 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkfzt\" (UniqueName: \"kubernetes.io/projected/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-kube-api-access-qkfzt\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.164930 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.178383 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.188372 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.269211 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkfzt\" (UniqueName: \"kubernetes.io/projected/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-kube-api-access-qkfzt\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.269269 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.273392 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.283572 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/838f0563-3ad1-4f68-9727-40140a0ba82a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.283657 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.283771 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-scripts\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.283845 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-ceph\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.283888 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.283931 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.284023 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-logs\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.284089 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838f0563-3ad1-4f68-9727-40140a0ba82a-logs\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.284154 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/838f0563-3ad1-4f68-9727-40140a0ba82a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.284179 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-config-data\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.284231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.284266 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx2gn\" (UniqueName: \"kubernetes.io/projected/838f0563-3ad1-4f68-9727-40140a0ba82a-kube-api-access-hx2gn\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.285947 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="513a2a30-d753-457f-b030-3a40a2a2f8eb" path="/var/lib/kubelet/pods/513a2a30-d753-457f-b030-3a40a2a2f8eb/volumes" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.287017 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6817c34-ecd6-4e01-8968-ad815733d564" path="/var/lib/kubelet/pods/a6817c34-ecd6-4e01-8968-ad815733d564/volumes" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.287907 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/838f0563-3ad1-4f68-9727-40140a0ba82a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.290669 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-logs\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.291775 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/838f0563-3ad1-4f68-9727-40140a0ba82a-logs\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.307227 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkfzt\" (UniqueName: \"kubernetes.io/projected/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-kube-api-access-qkfzt\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.307852 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/838f0563-3ad1-4f68-9727-40140a0ba82a-ceph\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.308531 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.308891 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-ceph\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.310553 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-config-data\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.325665 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.330942 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.331533 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b4a92f9a-86bc-45a1-9172-2fbc0e9be453-scripts\") pod \"glance-default-external-api-0\" (UID: \"b4a92f9a-86bc-45a1-9172-2fbc0e9be453\") " pod="openstack/glance-default-external-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.336915 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx2gn\" (UniqueName: \"kubernetes.io/projected/838f0563-3ad1-4f68-9727-40140a0ba82a-kube-api-access-hx2gn\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.350518 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/838f0563-3ad1-4f68-9727-40140a0ba82a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"838f0563-3ad1-4f68-9727-40140a0ba82a\") " pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.441190 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:22 crc kubenswrapper[4979]: I1204 13:23:22.467188 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 04 13:23:23 crc kubenswrapper[4979]: I1204 13:23:23.112031 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 04 13:23:23 crc kubenswrapper[4979]: W1204 13:23:23.122261 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod838f0563_3ad1_4f68_9727_40140a0ba82a.slice/crio-be811040f8ed55d4490db7dcb3c61c27b2b70abf7f3cfcc28c75235f39af4245 WatchSource:0}: Error finding container be811040f8ed55d4490db7dcb3c61c27b2b70abf7f3cfcc28c75235f39af4245: Status 404 returned error can't find the container with id be811040f8ed55d4490db7dcb3c61c27b2b70abf7f3cfcc28c75235f39af4245 Dec 04 13:23:23 crc kubenswrapper[4979]: I1204 13:23:23.227310 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 04 13:23:23 crc kubenswrapper[4979]: W1204 13:23:23.230583 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4a92f9a_86bc_45a1_9172_2fbc0e9be453.slice/crio-112b5b6eac01f59bcfc0098e1e04f80135f048c77cb85879a86e80126f70a055 WatchSource:0}: Error finding container 112b5b6eac01f59bcfc0098e1e04f80135f048c77cb85879a86e80126f70a055: Status 404 returned error can't find the container with id 112b5b6eac01f59bcfc0098e1e04f80135f048c77cb85879a86e80126f70a055 Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.006527 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4a92f9a-86bc-45a1-9172-2fbc0e9be453","Type":"ContainerStarted","Data":"f329acf83b7796d7c48d3b823c4aa4e11984e3fb502c08c3d66920ab0ce33462"} Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.007040 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4a92f9a-86bc-45a1-9172-2fbc0e9be453","Type":"ContainerStarted","Data":"112b5b6eac01f59bcfc0098e1e04f80135f048c77cb85879a86e80126f70a055"} Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.012387 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"838f0563-3ad1-4f68-9727-40140a0ba82a","Type":"ContainerStarted","Data":"90ee2be3ac93491b4020ef93e9ed05d106332985d6091f4d1e152017ab7e4148"} Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.012428 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"838f0563-3ad1-4f68-9727-40140a0ba82a","Type":"ContainerStarted","Data":"be811040f8ed55d4490db7dcb3c61c27b2b70abf7f3cfcc28c75235f39af4245"} Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.072859 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.215462 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.215520 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.801377 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:24 crc kubenswrapper[4979]: I1204 13:23:24.801699 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:25 crc kubenswrapper[4979]: I1204 13:23:25.026064 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"838f0563-3ad1-4f68-9727-40140a0ba82a","Type":"ContainerStarted","Data":"a968a38d9fb2973e148611def3717f404a654326c43cc0c4192db0c0d6cbbeb9"} Dec 04 13:23:25 crc kubenswrapper[4979]: I1204 13:23:25.031416 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b4a92f9a-86bc-45a1-9172-2fbc0e9be453","Type":"ContainerStarted","Data":"8f13085bff4c91a0753ab1d2b313f3123b3f77a4ea56db069b36a39fd16a09fa"} Dec 04 13:23:25 crc kubenswrapper[4979]: I1204 13:23:25.058984 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.058962229 podStartE2EDuration="3.058962229s" podCreationTimestamp="2025-12-04 13:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:23:25.052186699 +0000 UTC m=+6029.326482513" watchObservedRunningTime="2025-12-04 13:23:25.058962229 +0000 UTC m=+6029.333258033" Dec 04 13:23:25 crc kubenswrapper[4979]: I1204 13:23:25.075641 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.075623962 podStartE2EDuration="3.075623962s" podCreationTimestamp="2025-12-04 13:23:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:23:25.069890469 +0000 UTC m=+6029.344186293" watchObservedRunningTime="2025-12-04 13:23:25.075623962 +0000 UTC m=+6029.349919766" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.441831 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.443409 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.469085 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.469131 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.485916 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.502390 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.519726 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 13:23:32 crc kubenswrapper[4979]: I1204 13:23:32.519814 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 04 13:23:33 crc kubenswrapper[4979]: I1204 13:23:33.121730 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 13:23:33 crc kubenswrapper[4979]: I1204 13:23:33.122104 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:33 crc kubenswrapper[4979]: I1204 13:23:33.122123 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:33 crc kubenswrapper[4979]: I1204 13:23:33.122132 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 04 13:23:34 crc kubenswrapper[4979]: I1204 13:23:34.048808 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-mbnxj"] Dec 04 13:23:34 crc kubenswrapper[4979]: I1204 13:23:34.061858 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-mbnxj"] Dec 04 13:23:34 crc kubenswrapper[4979]: I1204 13:23:34.218159 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7b774f74bf-v9dzj" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 13:23:34 crc kubenswrapper[4979]: I1204 13:23:34.623238 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb5673d2-1830-410f-a8e5-ff1ff68ccd4b" path="/var/lib/kubelet/pods/bb5673d2-1830-410f-a8e5-ff1ff68ccd4b/volumes" Dec 04 13:23:34 crc kubenswrapper[4979]: I1204 13:23:34.803579 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-64f76457df-phmmr" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.037459 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-a773-account-create-update-t7h6h"] Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.073334 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-a773-account-create-update-t7h6h"] Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.795072 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.795197 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.801873 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.801984 4979 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.829737 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 04 13:23:35 crc kubenswrapper[4979]: I1204 13:23:35.832704 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 04 13:23:36 crc kubenswrapper[4979]: I1204 13:23:36.211299 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f887401-69a4-44e6-af00-97e2208c8137" path="/var/lib/kubelet/pods/7f887401-69a4-44e6-af00-97e2208c8137/volumes" Dec 04 13:23:45 crc kubenswrapper[4979]: I1204 13:23:45.031582 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-2crhq"] Dec 04 13:23:45 crc kubenswrapper[4979]: I1204 13:23:45.041193 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-2crhq"] Dec 04 13:23:46 crc kubenswrapper[4979]: I1204 13:23:46.149151 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:46 crc kubenswrapper[4979]: I1204 13:23:46.210117 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1174e2d1-da2d-4e5d-a1a0-9852a10457ee" path="/var/lib/kubelet/pods/1174e2d1-da2d-4e5d-a1a0-9852a10457ee/volumes" Dec 04 13:23:46 crc kubenswrapper[4979]: I1204 13:23:46.602185 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:47 crc kubenswrapper[4979]: I1204 13:23:47.857699 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:23:48 crc kubenswrapper[4979]: I1204 13:23:48.255767 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:23:48 crc kubenswrapper[4979]: I1204 13:23:48.321193 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b774f74bf-v9dzj"] Dec 04 13:23:48 crc kubenswrapper[4979]: I1204 13:23:48.321530 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b774f74bf-v9dzj" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" containerID="cri-o://811bf09226a25589eb9a35ebd7ffb031a18426a23b7b85e6c854ef2e40fd0fed" gracePeriod=30 Dec 04 13:23:48 crc kubenswrapper[4979]: I1204 13:23:48.321528 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b774f74bf-v9dzj" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon-log" containerID="cri-o://f830632c42ca4df4c807861d9a6e71f1057a520dda6a44e249bf4828d2a5bbb1" gracePeriod=30 Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.297044 4979 generic.go:334] "Generic (PLEG): container finished" podID="58966164-dd31-4e68-942e-97f6e9b78574" containerID="811bf09226a25589eb9a35ebd7ffb031a18426a23b7b85e6c854ef2e40fd0fed" exitCode=0 Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.297260 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b774f74bf-v9dzj" event={"ID":"58966164-dd31-4e68-942e-97f6e9b78574","Type":"ContainerDied","Data":"811bf09226a25589eb9a35ebd7ffb031a18426a23b7b85e6c854ef2e40fd0fed"} Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.300195 4979 generic.go:334] "Generic (PLEG): container finished" podID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerID="98a9830f316748de1b45673ea3faad97c57602e0787223babafd4e8d47228034" exitCode=137 Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.300248 4979 generic.go:334] "Generic (PLEG): container finished" podID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerID="22e6dbd3a9145842ea2090c72f861e332e0096c39dc9f53687a22e2321681a28" exitCode=137 Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.300276 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-554cf9495c-tbhnf" event={"ID":"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4","Type":"ContainerDied","Data":"98a9830f316748de1b45673ea3faad97c57602e0787223babafd4e8d47228034"} Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.300344 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-554cf9495c-tbhnf" event={"ID":"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4","Type":"ContainerDied","Data":"22e6dbd3a9145842ea2090c72f861e332e0096c39dc9f53687a22e2321681a28"} Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.300354 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-554cf9495c-tbhnf" event={"ID":"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4","Type":"ContainerDied","Data":"8966fcdabe19ce318304bbc3885c12a2579c7092608fda126190accbee60b561"} Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.300364 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8966fcdabe19ce318304bbc3885c12a2579c7092608fda126190accbee60b561" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.388548 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.433773 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-logs\") pod \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.433948 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-horizon-secret-key\") pod \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.433984 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-scripts\") pod \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.434258 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-logs" (OuterVolumeSpecName: "logs") pod "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" (UID: "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.434788 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-config-data\") pod \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.434862 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dk2q\" (UniqueName: \"kubernetes.io/projected/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-kube-api-access-5dk2q\") pod \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\" (UID: \"1c69ccc1-abcb-4368-8fb3-1b1725e89cc4\") " Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.435365 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.439922 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" (UID: "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.440222 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-kube-api-access-5dk2q" (OuterVolumeSpecName: "kube-api-access-5dk2q") pod "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" (UID: "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4"). InnerVolumeSpecName "kube-api-access-5dk2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.461435 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-scripts" (OuterVolumeSpecName: "scripts") pod "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" (UID: "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.475913 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-config-data" (OuterVolumeSpecName: "config-data") pod "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" (UID: "1c69ccc1-abcb-4368-8fb3-1b1725e89cc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.536867 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.536899 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dk2q\" (UniqueName: \"kubernetes.io/projected/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-kube-api-access-5dk2q\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.536910 4979 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:52 crc kubenswrapper[4979]: I1204 13:23:52.536919 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:23:53 crc kubenswrapper[4979]: I1204 13:23:53.307873 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-554cf9495c-tbhnf" Dec 04 13:23:53 crc kubenswrapper[4979]: I1204 13:23:53.342415 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-554cf9495c-tbhnf"] Dec 04 13:23:53 crc kubenswrapper[4979]: I1204 13:23:53.352438 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-554cf9495c-tbhnf"] Dec 04 13:23:54 crc kubenswrapper[4979]: I1204 13:23:54.210945 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" path="/var/lib/kubelet/pods/1c69ccc1-abcb-4368-8fb3-1b1725e89cc4/volumes" Dec 04 13:23:54 crc kubenswrapper[4979]: I1204 13:23:54.217126 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7b774f74bf-v9dzj" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.329591 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-788b9df87f-r5rlr"] Dec 04 13:23:56 crc kubenswrapper[4979]: E1204 13:23:56.330291 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.330318 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon" Dec 04 13:23:56 crc kubenswrapper[4979]: E1204 13:23:56.330370 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon-log" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.330376 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon-log" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.330565 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.330579 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c69ccc1-abcb-4368-8fb3-1b1725e89cc4" containerName="horizon-log" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.331610 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.349719 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-788b9df87f-r5rlr"] Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.417232 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a43dcb59-d096-43aa-8e38-465be32822e8-scripts\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.417292 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43dcb59-d096-43aa-8e38-465be32822e8-logs\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.417353 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zh5tz\" (UniqueName: \"kubernetes.io/projected/a43dcb59-d096-43aa-8e38-465be32822e8-kube-api-access-zh5tz\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.417590 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43dcb59-d096-43aa-8e38-465be32822e8-config-data\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.417758 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a43dcb59-d096-43aa-8e38-465be32822e8-horizon-secret-key\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.519736 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43dcb59-d096-43aa-8e38-465be32822e8-config-data\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.519829 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a43dcb59-d096-43aa-8e38-465be32822e8-horizon-secret-key\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.519955 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a43dcb59-d096-43aa-8e38-465be32822e8-scripts\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.520024 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43dcb59-d096-43aa-8e38-465be32822e8-logs\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.520068 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zh5tz\" (UniqueName: \"kubernetes.io/projected/a43dcb59-d096-43aa-8e38-465be32822e8-kube-api-access-zh5tz\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.520943 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a43dcb59-d096-43aa-8e38-465be32822e8-config-data\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.521467 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a43dcb59-d096-43aa-8e38-465be32822e8-scripts\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.521657 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a43dcb59-d096-43aa-8e38-465be32822e8-logs\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.527468 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/a43dcb59-d096-43aa-8e38-465be32822e8-horizon-secret-key\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.543498 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zh5tz\" (UniqueName: \"kubernetes.io/projected/a43dcb59-d096-43aa-8e38-465be32822e8-kube-api-access-zh5tz\") pod \"horizon-788b9df87f-r5rlr\" (UID: \"a43dcb59-d096-43aa-8e38-465be32822e8\") " pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:56 crc kubenswrapper[4979]: I1204 13:23:56.655127 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.124362 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-788b9df87f-r5rlr"] Dec 04 13:23:57 crc kubenswrapper[4979]: W1204 13:23:57.128136 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda43dcb59_d096_43aa_8e38_465be32822e8.slice/crio-01e19ac7707ac5387efc7d0bd9e07428d5fea1c7bb496fbe7d064adf372643fb WatchSource:0}: Error finding container 01e19ac7707ac5387efc7d0bd9e07428d5fea1c7bb496fbe7d064adf372643fb: Status 404 returned error can't find the container with id 01e19ac7707ac5387efc7d0bd9e07428d5fea1c7bb496fbe7d064adf372643fb Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.350899 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788b9df87f-r5rlr" event={"ID":"a43dcb59-d096-43aa-8e38-465be32822e8","Type":"ContainerStarted","Data":"3fce9db2a46924796e61e960a1f0861c53d4b965bc48844ab335385063d148b6"} Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.352280 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788b9df87f-r5rlr" event={"ID":"a43dcb59-d096-43aa-8e38-465be32822e8","Type":"ContainerStarted","Data":"01e19ac7707ac5387efc7d0bd9e07428d5fea1c7bb496fbe7d064adf372643fb"} Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.637789 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-2czth"] Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.639534 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.641156 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/055c74a7-df67-4434-960e-e0deab4e5827-operator-scripts\") pod \"heat-db-create-2czth\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.641280 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7h86\" (UniqueName: \"kubernetes.io/projected/055c74a7-df67-4434-960e-e0deab4e5827-kube-api-access-k7h86\") pod \"heat-db-create-2czth\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.650713 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-6ad5-account-create-update-7p225"] Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.652049 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.658900 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.661948 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-2czth"] Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.677620 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6ad5-account-create-update-7p225"] Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.742831 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/055c74a7-df67-4434-960e-e0deab4e5827-operator-scripts\") pod \"heat-db-create-2czth\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.742961 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7h86\" (UniqueName: \"kubernetes.io/projected/055c74a7-df67-4434-960e-e0deab4e5827-kube-api-access-k7h86\") pod \"heat-db-create-2czth\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.743860 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/055c74a7-df67-4434-960e-e0deab4e5827-operator-scripts\") pod \"heat-db-create-2czth\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.760028 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7h86\" (UniqueName: \"kubernetes.io/projected/055c74a7-df67-4434-960e-e0deab4e5827-kube-api-access-k7h86\") pod \"heat-db-create-2czth\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.845051 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47eb1d7d-1fe4-4491-8584-b667a830621a-operator-scripts\") pod \"heat-6ad5-account-create-update-7p225\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.845653 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbqf\" (UniqueName: \"kubernetes.io/projected/47eb1d7d-1fe4-4491-8584-b667a830621a-kube-api-access-4fbqf\") pod \"heat-6ad5-account-create-update-7p225\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.947519 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47eb1d7d-1fe4-4491-8584-b667a830621a-operator-scripts\") pod \"heat-6ad5-account-create-update-7p225\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.947652 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbqf\" (UniqueName: \"kubernetes.io/projected/47eb1d7d-1fe4-4491-8584-b667a830621a-kube-api-access-4fbqf\") pod \"heat-6ad5-account-create-update-7p225\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.948280 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47eb1d7d-1fe4-4491-8584-b667a830621a-operator-scripts\") pod \"heat-6ad5-account-create-update-7p225\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.963353 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbqf\" (UniqueName: \"kubernetes.io/projected/47eb1d7d-1fe4-4491-8584-b667a830621a-kube-api-access-4fbqf\") pod \"heat-6ad5-account-create-update-7p225\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.964467 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2czth" Dec 04 13:23:57 crc kubenswrapper[4979]: I1204 13:23:57.985088 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:23:58 crc kubenswrapper[4979]: I1204 13:23:58.365147 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-788b9df87f-r5rlr" event={"ID":"a43dcb59-d096-43aa-8e38-465be32822e8","Type":"ContainerStarted","Data":"e80654d50b7a05fb92828e4cbeed43501d4f61915ace356d272d3056f0bb7b20"} Dec 04 13:23:58 crc kubenswrapper[4979]: I1204 13:23:58.399423 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-788b9df87f-r5rlr" podStartSLOduration=2.399397231 podStartE2EDuration="2.399397231s" podCreationTimestamp="2025-12-04 13:23:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:23:58.387910904 +0000 UTC m=+6062.662206708" watchObservedRunningTime="2025-12-04 13:23:58.399397231 +0000 UTC m=+6062.673693055" Dec 04 13:23:58 crc kubenswrapper[4979]: W1204 13:23:58.460992 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod055c74a7_df67_4434_960e_e0deab4e5827.slice/crio-382f2a36dbfb4e1942c6bafba90651bb9fcbd7a376dddd065c057c14d56048b0 WatchSource:0}: Error finding container 382f2a36dbfb4e1942c6bafba90651bb9fcbd7a376dddd065c057c14d56048b0: Status 404 returned error can't find the container with id 382f2a36dbfb4e1942c6bafba90651bb9fcbd7a376dddd065c057c14d56048b0 Dec 04 13:23:58 crc kubenswrapper[4979]: I1204 13:23:58.458727 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-2czth"] Dec 04 13:23:58 crc kubenswrapper[4979]: I1204 13:23:58.537044 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-6ad5-account-create-update-7p225"] Dec 04 13:23:59 crc kubenswrapper[4979]: I1204 13:23:59.375714 4979 generic.go:334] "Generic (PLEG): container finished" podID="055c74a7-df67-4434-960e-e0deab4e5827" containerID="299093ef32438cc3517c379c1ee45df6194fe9f79c468445dcee7a57fb7bf1ed" exitCode=0 Dec 04 13:23:59 crc kubenswrapper[4979]: I1204 13:23:59.375798 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2czth" event={"ID":"055c74a7-df67-4434-960e-e0deab4e5827","Type":"ContainerDied","Data":"299093ef32438cc3517c379c1ee45df6194fe9f79c468445dcee7a57fb7bf1ed"} Dec 04 13:23:59 crc kubenswrapper[4979]: I1204 13:23:59.376368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2czth" event={"ID":"055c74a7-df67-4434-960e-e0deab4e5827","Type":"ContainerStarted","Data":"382f2a36dbfb4e1942c6bafba90651bb9fcbd7a376dddd065c057c14d56048b0"} Dec 04 13:23:59 crc kubenswrapper[4979]: I1204 13:23:59.377509 4979 generic.go:334] "Generic (PLEG): container finished" podID="47eb1d7d-1fe4-4491-8584-b667a830621a" containerID="fa84032727b1245143fdb17593f307cbc0738908e870c1446d6cd3e0a52425d4" exitCode=0 Dec 04 13:23:59 crc kubenswrapper[4979]: I1204 13:23:59.379119 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6ad5-account-create-update-7p225" event={"ID":"47eb1d7d-1fe4-4491-8584-b667a830621a","Type":"ContainerDied","Data":"fa84032727b1245143fdb17593f307cbc0738908e870c1446d6cd3e0a52425d4"} Dec 04 13:23:59 crc kubenswrapper[4979]: I1204 13:23:59.379149 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6ad5-account-create-update-7p225" event={"ID":"47eb1d7d-1fe4-4491-8584-b667a830621a","Type":"ContainerStarted","Data":"dce32283c700919c1c461c6fa8a8c212f8d0cf96037e2dbae10765d30a1bb023"} Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.803903 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2czth" Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.808279 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.909128 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fbqf\" (UniqueName: \"kubernetes.io/projected/47eb1d7d-1fe4-4491-8584-b667a830621a-kube-api-access-4fbqf\") pod \"47eb1d7d-1fe4-4491-8584-b667a830621a\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.909355 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47eb1d7d-1fe4-4491-8584-b667a830621a-operator-scripts\") pod \"47eb1d7d-1fe4-4491-8584-b667a830621a\" (UID: \"47eb1d7d-1fe4-4491-8584-b667a830621a\") " Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.909390 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/055c74a7-df67-4434-960e-e0deab4e5827-operator-scripts\") pod \"055c74a7-df67-4434-960e-e0deab4e5827\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.909423 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7h86\" (UniqueName: \"kubernetes.io/projected/055c74a7-df67-4434-960e-e0deab4e5827-kube-api-access-k7h86\") pod \"055c74a7-df67-4434-960e-e0deab4e5827\" (UID: \"055c74a7-df67-4434-960e-e0deab4e5827\") " Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.911383 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/055c74a7-df67-4434-960e-e0deab4e5827-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "055c74a7-df67-4434-960e-e0deab4e5827" (UID: "055c74a7-df67-4434-960e-e0deab4e5827"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.911835 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47eb1d7d-1fe4-4491-8584-b667a830621a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47eb1d7d-1fe4-4491-8584-b667a830621a" (UID: "47eb1d7d-1fe4-4491-8584-b667a830621a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.916014 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47eb1d7d-1fe4-4491-8584-b667a830621a-kube-api-access-4fbqf" (OuterVolumeSpecName: "kube-api-access-4fbqf") pod "47eb1d7d-1fe4-4491-8584-b667a830621a" (UID: "47eb1d7d-1fe4-4491-8584-b667a830621a"). InnerVolumeSpecName "kube-api-access-4fbqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:24:00 crc kubenswrapper[4979]: I1204 13:24:00.916328 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/055c74a7-df67-4434-960e-e0deab4e5827-kube-api-access-k7h86" (OuterVolumeSpecName: "kube-api-access-k7h86") pod "055c74a7-df67-4434-960e-e0deab4e5827" (UID: "055c74a7-df67-4434-960e-e0deab4e5827"). InnerVolumeSpecName "kube-api-access-k7h86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.011763 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47eb1d7d-1fe4-4491-8584-b667a830621a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.011792 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/055c74a7-df67-4434-960e-e0deab4e5827-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.011805 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7h86\" (UniqueName: \"kubernetes.io/projected/055c74a7-df67-4434-960e-e0deab4e5827-kube-api-access-k7h86\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.011816 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fbqf\" (UniqueName: \"kubernetes.io/projected/47eb1d7d-1fe4-4491-8584-b667a830621a-kube-api-access-4fbqf\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.398158 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-2czth" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.398153 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-2czth" event={"ID":"055c74a7-df67-4434-960e-e0deab4e5827","Type":"ContainerDied","Data":"382f2a36dbfb4e1942c6bafba90651bb9fcbd7a376dddd065c057c14d56048b0"} Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.398273 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="382f2a36dbfb4e1942c6bafba90651bb9fcbd7a376dddd065c057c14d56048b0" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.400993 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-6ad5-account-create-update-7p225" event={"ID":"47eb1d7d-1fe4-4491-8584-b667a830621a","Type":"ContainerDied","Data":"dce32283c700919c1c461c6fa8a8c212f8d0cf96037e2dbae10765d30a1bb023"} Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.401022 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dce32283c700919c1c461c6fa8a8c212f8d0cf96037e2dbae10765d30a1bb023" Dec 04 13:24:01 crc kubenswrapper[4979]: I1204 13:24:01.401063 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-6ad5-account-create-update-7p225" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.881149 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-nvnkz"] Dec 04 13:24:02 crc kubenswrapper[4979]: E1204 13:24:02.881869 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="055c74a7-df67-4434-960e-e0deab4e5827" containerName="mariadb-database-create" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.881885 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="055c74a7-df67-4434-960e-e0deab4e5827" containerName="mariadb-database-create" Dec 04 13:24:02 crc kubenswrapper[4979]: E1204 13:24:02.881911 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47eb1d7d-1fe4-4491-8584-b667a830621a" containerName="mariadb-account-create-update" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.881917 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="47eb1d7d-1fe4-4491-8584-b667a830621a" containerName="mariadb-account-create-update" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.882106 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="47eb1d7d-1fe4-4491-8584-b667a830621a" containerName="mariadb-account-create-update" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.882130 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="055c74a7-df67-4434-960e-e0deab4e5827" containerName="mariadb-database-create" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.882821 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.884723 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-xxctp" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.885900 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 04 13:24:02 crc kubenswrapper[4979]: I1204 13:24:02.894076 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-nvnkz"] Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.050745 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd6fl\" (UniqueName: \"kubernetes.io/projected/b321d456-d89e-45f2-aa4a-807af0bc5028-kube-api-access-bd6fl\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.050848 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-config-data\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.050900 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-combined-ca-bundle\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.153026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-combined-ca-bundle\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.153218 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd6fl\" (UniqueName: \"kubernetes.io/projected/b321d456-d89e-45f2-aa4a-807af0bc5028-kube-api-access-bd6fl\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.153290 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-config-data\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.167349 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-combined-ca-bundle\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.168991 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-config-data\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.169995 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd6fl\" (UniqueName: \"kubernetes.io/projected/b321d456-d89e-45f2-aa4a-807af0bc5028-kube-api-access-bd6fl\") pod \"heat-db-sync-nvnkz\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.200128 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:03 crc kubenswrapper[4979]: I1204 13:24:03.634062 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-nvnkz"] Dec 04 13:24:03 crc kubenswrapper[4979]: W1204 13:24:03.637728 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb321d456_d89e_45f2_aa4a_807af0bc5028.slice/crio-bb412198be4599f90984bb79b7401c008dd2dd66f198ab204e3352a680cc889f WatchSource:0}: Error finding container bb412198be4599f90984bb79b7401c008dd2dd66f198ab204e3352a680cc889f: Status 404 returned error can't find the container with id bb412198be4599f90984bb79b7401c008dd2dd66f198ab204e3352a680cc889f Dec 04 13:24:04 crc kubenswrapper[4979]: I1204 13:24:04.216247 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7b774f74bf-v9dzj" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 13:24:04 crc kubenswrapper[4979]: I1204 13:24:04.430520 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nvnkz" event={"ID":"b321d456-d89e-45f2-aa4a-807af0bc5028","Type":"ContainerStarted","Data":"bb412198be4599f90984bb79b7401c008dd2dd66f198ab204e3352a680cc889f"} Dec 04 13:24:06 crc kubenswrapper[4979]: I1204 13:24:06.655949 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:24:06 crc kubenswrapper[4979]: I1204 13:24:06.656274 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:24:08 crc kubenswrapper[4979]: I1204 13:24:08.260215 4979 scope.go:117] "RemoveContainer" containerID="0b7e96c223001574ccd4e65dc7936441d4f164b47fee37c67117d33eb723e09a" Dec 04 13:24:09 crc kubenswrapper[4979]: I1204 13:24:09.533493 4979 scope.go:117] "RemoveContainer" containerID="76af69893ea3fc1b963f5f63db2c5133081472afa4514a0861b2318ba91ad508" Dec 04 13:24:09 crc kubenswrapper[4979]: I1204 13:24:09.793418 4979 scope.go:117] "RemoveContainer" containerID="2d272ff8edba3842c49c6a817a8ab18970d87dc215ec67a73bb759a47cc5b909" Dec 04 13:24:10 crc kubenswrapper[4979]: I1204 13:24:10.497951 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nvnkz" event={"ID":"b321d456-d89e-45f2-aa4a-807af0bc5028","Type":"ContainerStarted","Data":"2b2fc19afe3e94939e6780da1a95b4aee8d4904f5e72c45ca9885074347fa37d"} Dec 04 13:24:10 crc kubenswrapper[4979]: I1204 13:24:10.525828 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-nvnkz" podStartSLOduration=2.312492531 podStartE2EDuration="8.525806424s" podCreationTimestamp="2025-12-04 13:24:02 +0000 UTC" firstStartedPulling="2025-12-04 13:24:03.640648012 +0000 UTC m=+6067.914943816" lastFinishedPulling="2025-12-04 13:24:09.853961905 +0000 UTC m=+6074.128257709" observedRunningTime="2025-12-04 13:24:10.517408081 +0000 UTC m=+6074.791703905" watchObservedRunningTime="2025-12-04 13:24:10.525806424 +0000 UTC m=+6074.800102228" Dec 04 13:24:12 crc kubenswrapper[4979]: I1204 13:24:12.519490 4979 generic.go:334] "Generic (PLEG): container finished" podID="b321d456-d89e-45f2-aa4a-807af0bc5028" containerID="2b2fc19afe3e94939e6780da1a95b4aee8d4904f5e72c45ca9885074347fa37d" exitCode=0 Dec 04 13:24:12 crc kubenswrapper[4979]: I1204 13:24:12.519666 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nvnkz" event={"ID":"b321d456-d89e-45f2-aa4a-807af0bc5028","Type":"ContainerDied","Data":"2b2fc19afe3e94939e6780da1a95b4aee8d4904f5e72c45ca9885074347fa37d"} Dec 04 13:24:13 crc kubenswrapper[4979]: I1204 13:24:13.879123 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:13 crc kubenswrapper[4979]: I1204 13:24:13.971621 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd6fl\" (UniqueName: \"kubernetes.io/projected/b321d456-d89e-45f2-aa4a-807af0bc5028-kube-api-access-bd6fl\") pod \"b321d456-d89e-45f2-aa4a-807af0bc5028\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " Dec 04 13:24:13 crc kubenswrapper[4979]: I1204 13:24:13.972023 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-combined-ca-bundle\") pod \"b321d456-d89e-45f2-aa4a-807af0bc5028\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " Dec 04 13:24:13 crc kubenswrapper[4979]: I1204 13:24:13.972067 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-config-data\") pod \"b321d456-d89e-45f2-aa4a-807af0bc5028\" (UID: \"b321d456-d89e-45f2-aa4a-807af0bc5028\") " Dec 04 13:24:13 crc kubenswrapper[4979]: I1204 13:24:13.980180 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b321d456-d89e-45f2-aa4a-807af0bc5028-kube-api-access-bd6fl" (OuterVolumeSpecName: "kube-api-access-bd6fl") pod "b321d456-d89e-45f2-aa4a-807af0bc5028" (UID: "b321d456-d89e-45f2-aa4a-807af0bc5028"). InnerVolumeSpecName "kube-api-access-bd6fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.000151 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b321d456-d89e-45f2-aa4a-807af0bc5028" (UID: "b321d456-d89e-45f2-aa4a-807af0bc5028"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.045245 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-config-data" (OuterVolumeSpecName: "config-data") pod "b321d456-d89e-45f2-aa4a-807af0bc5028" (UID: "b321d456-d89e-45f2-aa4a-807af0bc5028"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.077581 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd6fl\" (UniqueName: \"kubernetes.io/projected/b321d456-d89e-45f2-aa4a-807af0bc5028-kube-api-access-bd6fl\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.077626 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.077637 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b321d456-d89e-45f2-aa4a-807af0bc5028-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.216479 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7b774f74bf-v9dzj" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.115:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.115:8080: connect: connection refused" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.216775 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.552167 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-nvnkz" event={"ID":"b321d456-d89e-45f2-aa4a-807af0bc5028","Type":"ContainerDied","Data":"bb412198be4599f90984bb79b7401c008dd2dd66f198ab204e3352a680cc889f"} Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.552959 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb412198be4599f90984bb79b7401c008dd2dd66f198ab204e3352a680cc889f" Dec 04 13:24:14 crc kubenswrapper[4979]: I1204 13:24:14.552231 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-nvnkz" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.543264 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7fd6595c6b-wn2bx"] Dec 04 13:24:15 crc kubenswrapper[4979]: E1204 13:24:15.544009 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b321d456-d89e-45f2-aa4a-807af0bc5028" containerName="heat-db-sync" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.544022 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b321d456-d89e-45f2-aa4a-807af0bc5028" containerName="heat-db-sync" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.544223 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b321d456-d89e-45f2-aa4a-807af0bc5028" containerName="heat-db-sync" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.544896 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.548525 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.549122 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-xxctp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.549256 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.561663 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fd6595c6b-wn2bx"] Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.665628 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-77ff7f9667-pk7gp"] Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.666894 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.676503 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.682397 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77ff7f9667-pk7gp"] Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.704818 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-68b7f67857-gnjfl"] Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.706182 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.708146 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.713073 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-config-data-custom\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.713430 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxjfk\" (UniqueName: \"kubernetes.io/projected/2628d26a-412b-43da-8525-b5375ea17e83-kube-api-access-zxjfk\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.713490 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-config-data\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.713562 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-combined-ca-bundle\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.716516 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68b7f67857-gnjfl"] Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815330 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-combined-ca-bundle\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815440 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-config-data-custom\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815470 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6f9z\" (UniqueName: \"kubernetes.io/projected/074b9676-de98-460f-9979-2790fc9778f5-kube-api-access-q6f9z\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815512 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-config-data-custom\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815565 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-config-data\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815594 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxjjx\" (UniqueName: \"kubernetes.io/projected/28eff767-0937-4b93-950d-d35c099e3ce5-kube-api-access-lxjjx\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815636 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-config-data-custom\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815673 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-combined-ca-bundle\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815722 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxjfk\" (UniqueName: \"kubernetes.io/projected/2628d26a-412b-43da-8525-b5375ea17e83-kube-api-access-zxjfk\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815752 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-combined-ca-bundle\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815782 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-config-data\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.815810 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-config-data\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.821374 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-combined-ca-bundle\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.822843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-config-data\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.824199 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2628d26a-412b-43da-8525-b5375ea17e83-config-data-custom\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.837177 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxjfk\" (UniqueName: \"kubernetes.io/projected/2628d26a-412b-43da-8525-b5375ea17e83-kube-api-access-zxjfk\") pod \"heat-engine-7fd6595c6b-wn2bx\" (UID: \"2628d26a-412b-43da-8525-b5375ea17e83\") " pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.897812 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917614 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-config-data-custom\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917672 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-combined-ca-bundle\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917720 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-combined-ca-bundle\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917738 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-config-data\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917831 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6f9z\" (UniqueName: \"kubernetes.io/projected/074b9676-de98-460f-9979-2790fc9778f5-kube-api-access-q6f9z\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917889 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-config-data-custom\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917928 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-config-data\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.917946 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxjjx\" (UniqueName: \"kubernetes.io/projected/28eff767-0937-4b93-950d-d35c099e3ce5-kube-api-access-lxjjx\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.926058 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-combined-ca-bundle\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.926489 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-config-data\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.927526 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-config-data\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.938349 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-config-data-custom\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.941194 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/074b9676-de98-460f-9979-2790fc9778f5-combined-ca-bundle\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.941278 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/28eff767-0937-4b93-950d-d35c099e3ce5-config-data-custom\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.960403 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxjjx\" (UniqueName: \"kubernetes.io/projected/28eff767-0937-4b93-950d-d35c099e3ce5-kube-api-access-lxjjx\") pod \"heat-api-77ff7f9667-pk7gp\" (UID: \"28eff767-0937-4b93-950d-d35c099e3ce5\") " pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.980043 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6f9z\" (UniqueName: \"kubernetes.io/projected/074b9676-de98-460f-9979-2790fc9778f5-kube-api-access-q6f9z\") pod \"heat-cfnapi-68b7f67857-gnjfl\" (UID: \"074b9676-de98-460f-9979-2790fc9778f5\") " pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:15 crc kubenswrapper[4979]: I1204 13:24:15.990889 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:16 crc kubenswrapper[4979]: I1204 13:24:16.039017 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:16 crc kubenswrapper[4979]: W1204 13:24:16.560847 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28eff767_0937_4b93_950d_d35c099e3ce5.slice/crio-d61e97ea3427b048fd21ea264b63009aa719e51069631de3c723163bd4dbecac WatchSource:0}: Error finding container d61e97ea3427b048fd21ea264b63009aa719e51069631de3c723163bd4dbecac: Status 404 returned error can't find the container with id d61e97ea3427b048fd21ea264b63009aa719e51069631de3c723163bd4dbecac Dec 04 13:24:16 crc kubenswrapper[4979]: I1204 13:24:16.569499 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77ff7f9667-pk7gp"] Dec 04 13:24:16 crc kubenswrapper[4979]: I1204 13:24:16.616587 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77ff7f9667-pk7gp" event={"ID":"28eff767-0937-4b93-950d-d35c099e3ce5","Type":"ContainerStarted","Data":"d61e97ea3427b048fd21ea264b63009aa719e51069631de3c723163bd4dbecac"} Dec 04 13:24:16 crc kubenswrapper[4979]: I1204 13:24:16.661543 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-788b9df87f-r5rlr" podUID="a43dcb59-d096-43aa-8e38-465be32822e8" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.119:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.119:8080: connect: connection refused" Dec 04 13:24:16 crc kubenswrapper[4979]: W1204 13:24:16.864661 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2628d26a_412b_43da_8525_b5375ea17e83.slice/crio-45328d32aaf55306c79d3f123a2af47ec97bc80b40a26938c6ecba70e77c010f WatchSource:0}: Error finding container 45328d32aaf55306c79d3f123a2af47ec97bc80b40a26938c6ecba70e77c010f: Status 404 returned error can't find the container with id 45328d32aaf55306c79d3f123a2af47ec97bc80b40a26938c6ecba70e77c010f Dec 04 13:24:16 crc kubenswrapper[4979]: I1204 13:24:16.892689 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-68b7f67857-gnjfl"] Dec 04 13:24:16 crc kubenswrapper[4979]: I1204 13:24:16.913339 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7fd6595c6b-wn2bx"] Dec 04 13:24:17 crc kubenswrapper[4979]: I1204 13:24:17.632128 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fd6595c6b-wn2bx" event={"ID":"2628d26a-412b-43da-8525-b5375ea17e83","Type":"ContainerStarted","Data":"f3403882735f6ac92586db93bca4414ca457efd672b46c3c09cd893056bf583e"} Dec 04 13:24:17 crc kubenswrapper[4979]: I1204 13:24:17.632789 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7fd6595c6b-wn2bx" event={"ID":"2628d26a-412b-43da-8525-b5375ea17e83","Type":"ContainerStarted","Data":"45328d32aaf55306c79d3f123a2af47ec97bc80b40a26938c6ecba70e77c010f"} Dec 04 13:24:17 crc kubenswrapper[4979]: I1204 13:24:17.632811 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:17 crc kubenswrapper[4979]: I1204 13:24:17.633900 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" event={"ID":"074b9676-de98-460f-9979-2790fc9778f5","Type":"ContainerStarted","Data":"028f58d13d30042ff46f7d9ea60ee296836e008e1b3ae1ee0428e0e9e7b4da83"} Dec 04 13:24:17 crc kubenswrapper[4979]: I1204 13:24:17.659466 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7fd6595c6b-wn2bx" podStartSLOduration=2.659442937 podStartE2EDuration="2.659442937s" podCreationTimestamp="2025-12-04 13:24:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:24:17.647842706 +0000 UTC m=+6081.922138510" watchObservedRunningTime="2025-12-04 13:24:17.659442937 +0000 UTC m=+6081.933738741" Dec 04 13:24:18 crc kubenswrapper[4979]: I1204 13:24:18.658361 4979 generic.go:334] "Generic (PLEG): container finished" podID="58966164-dd31-4e68-942e-97f6e9b78574" containerID="f830632c42ca4df4c807861d9a6e71f1057a520dda6a44e249bf4828d2a5bbb1" exitCode=137 Dec 04 13:24:18 crc kubenswrapper[4979]: I1204 13:24:18.658822 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b774f74bf-v9dzj" event={"ID":"58966164-dd31-4e68-942e-97f6e9b78574","Type":"ContainerDied","Data":"f830632c42ca4df4c807861d9a6e71f1057a520dda6a44e249bf4828d2a5bbb1"} Dec 04 13:24:18 crc kubenswrapper[4979]: I1204 13:24:18.661412 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77ff7f9667-pk7gp" event={"ID":"28eff767-0937-4b93-950d-d35c099e3ce5","Type":"ContainerStarted","Data":"3598e535db54f4d86a9d541d5273fe9f4404307532d2c328c5c5c5b5cf8f72f9"} Dec 04 13:24:18 crc kubenswrapper[4979]: I1204 13:24:18.661470 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:18 crc kubenswrapper[4979]: I1204 13:24:18.679519 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-77ff7f9667-pk7gp" podStartSLOduration=2.233116556 podStartE2EDuration="3.67950252s" podCreationTimestamp="2025-12-04 13:24:15 +0000 UTC" firstStartedPulling="2025-12-04 13:24:16.569412019 +0000 UTC m=+6080.843707813" lastFinishedPulling="2025-12-04 13:24:18.015797973 +0000 UTC m=+6082.290093777" observedRunningTime="2025-12-04 13:24:18.677166667 +0000 UTC m=+6082.951462481" watchObservedRunningTime="2025-12-04 13:24:18.67950252 +0000 UTC m=+6082.953798314" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.362532 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.495093 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-config-data\") pod \"58966164-dd31-4e68-942e-97f6e9b78574\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.495190 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k72p\" (UniqueName: \"kubernetes.io/projected/58966164-dd31-4e68-942e-97f6e9b78574-kube-api-access-5k72p\") pod \"58966164-dd31-4e68-942e-97f6e9b78574\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.495246 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/58966164-dd31-4e68-942e-97f6e9b78574-horizon-secret-key\") pod \"58966164-dd31-4e68-942e-97f6e9b78574\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.495318 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-scripts\") pod \"58966164-dd31-4e68-942e-97f6e9b78574\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.495343 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58966164-dd31-4e68-942e-97f6e9b78574-logs\") pod \"58966164-dd31-4e68-942e-97f6e9b78574\" (UID: \"58966164-dd31-4e68-942e-97f6e9b78574\") " Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.496349 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58966164-dd31-4e68-942e-97f6e9b78574-logs" (OuterVolumeSpecName: "logs") pod "58966164-dd31-4e68-942e-97f6e9b78574" (UID: "58966164-dd31-4e68-942e-97f6e9b78574"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.577922 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58966164-dd31-4e68-942e-97f6e9b78574-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "58966164-dd31-4e68-942e-97f6e9b78574" (UID: "58966164-dd31-4e68-942e-97f6e9b78574"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.578160 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58966164-dd31-4e68-942e-97f6e9b78574-kube-api-access-5k72p" (OuterVolumeSpecName: "kube-api-access-5k72p") pod "58966164-dd31-4e68-942e-97f6e9b78574" (UID: "58966164-dd31-4e68-942e-97f6e9b78574"). InnerVolumeSpecName "kube-api-access-5k72p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.584100 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-scripts" (OuterVolumeSpecName: "scripts") pod "58966164-dd31-4e68-942e-97f6e9b78574" (UID: "58966164-dd31-4e68-942e-97f6e9b78574"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.598462 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k72p\" (UniqueName: \"kubernetes.io/projected/58966164-dd31-4e68-942e-97f6e9b78574-kube-api-access-5k72p\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.598503 4979 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/58966164-dd31-4e68-942e-97f6e9b78574-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.598517 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.598528 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58966164-dd31-4e68-942e-97f6e9b78574-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.621048 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-config-data" (OuterVolumeSpecName: "config-data") pod "58966164-dd31-4e68-942e-97f6e9b78574" (UID: "58966164-dd31-4e68-942e-97f6e9b78574"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.674344 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b774f74bf-v9dzj" event={"ID":"58966164-dd31-4e68-942e-97f6e9b78574","Type":"ContainerDied","Data":"9ec5ba78b907fd11744236267b125cdbe4a624537a901c0e6e14f497198a7274"} Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.674392 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b774f74bf-v9dzj" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.674871 4979 scope.go:117] "RemoveContainer" containerID="811bf09226a25589eb9a35ebd7ffb031a18426a23b7b85e6c854ef2e40fd0fed" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.705908 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/58966164-dd31-4e68-942e-97f6e9b78574-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.710722 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b774f74bf-v9dzj"] Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.720762 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b774f74bf-v9dzj"] Dec 04 13:24:19 crc kubenswrapper[4979]: I1204 13:24:19.980417 4979 scope.go:117] "RemoveContainer" containerID="f830632c42ca4df4c807861d9a6e71f1057a520dda6a44e249bf4828d2a5bbb1" Dec 04 13:24:20 crc kubenswrapper[4979]: I1204 13:24:20.211334 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58966164-dd31-4e68-942e-97f6e9b78574" path="/var/lib/kubelet/pods/58966164-dd31-4e68-942e-97f6e9b78574/volumes" Dec 04 13:24:20 crc kubenswrapper[4979]: I1204 13:24:20.683969 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" event={"ID":"074b9676-de98-460f-9979-2790fc9778f5","Type":"ContainerStarted","Data":"7c00b44c114d016f17e12f8a1e6d005bc0f856f4790dc14950c5ba58e24aec1d"} Dec 04 13:24:20 crc kubenswrapper[4979]: I1204 13:24:20.685266 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:20 crc kubenswrapper[4979]: I1204 13:24:20.706199 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" podStartSLOduration=3.38901773 podStartE2EDuration="5.706181393s" podCreationTimestamp="2025-12-04 13:24:15 +0000 UTC" firstStartedPulling="2025-12-04 13:24:16.870259896 +0000 UTC m=+6081.144555710" lastFinishedPulling="2025-12-04 13:24:19.187423569 +0000 UTC m=+6083.461719373" observedRunningTime="2025-12-04 13:24:20.699130034 +0000 UTC m=+6084.973425838" watchObservedRunningTime="2025-12-04 13:24:20.706181393 +0000 UTC m=+6084.980477197" Dec 04 13:24:27 crc kubenswrapper[4979]: I1204 13:24:27.064840 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-e2de-account-create-update-gmnz9"] Dec 04 13:24:27 crc kubenswrapper[4979]: I1204 13:24:27.076970 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-6j2nx"] Dec 04 13:24:27 crc kubenswrapper[4979]: I1204 13:24:27.090908 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-6j2nx"] Dec 04 13:24:27 crc kubenswrapper[4979]: I1204 13:24:27.104879 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-e2de-account-create-update-gmnz9"] Dec 04 13:24:27 crc kubenswrapper[4979]: I1204 13:24:27.323135 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-77ff7f9667-pk7gp" Dec 04 13:24:27 crc kubenswrapper[4979]: I1204 13:24:27.464725 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-68b7f67857-gnjfl" Dec 04 13:24:28 crc kubenswrapper[4979]: I1204 13:24:28.211216 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="074dff34-e984-42cf-a467-ac5e96716fec" path="/var/lib/kubelet/pods/074dff34-e984-42cf-a467-ac5e96716fec/volumes" Dec 04 13:24:28 crc kubenswrapper[4979]: I1204 13:24:28.213142 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="540d2569-0399-40eb-a451-e43cf58270bc" path="/var/lib/kubelet/pods/540d2569-0399-40eb-a451-e43cf58270bc/volumes" Dec 04 13:24:28 crc kubenswrapper[4979]: I1204 13:24:28.668359 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:24:30 crc kubenswrapper[4979]: I1204 13:24:30.516837 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-788b9df87f-r5rlr" Dec 04 13:24:30 crc kubenswrapper[4979]: I1204 13:24:30.584543 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64f76457df-phmmr"] Dec 04 13:24:30 crc kubenswrapper[4979]: I1204 13:24:30.584788 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64f76457df-phmmr" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon-log" containerID="cri-o://19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8" gracePeriod=30 Dec 04 13:24:30 crc kubenswrapper[4979]: I1204 13:24:30.585186 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-64f76457df-phmmr" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" containerID="cri-o://f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed" gracePeriod=30 Dec 04 13:24:34 crc kubenswrapper[4979]: I1204 13:24:34.801490 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f76457df-phmmr" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Dec 04 13:24:34 crc kubenswrapper[4979]: I1204 13:24:34.832364 4979 generic.go:334] "Generic (PLEG): container finished" podID="016a470f-300c-46b0-9241-f73f843de724" containerID="f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed" exitCode=0 Dec 04 13:24:34 crc kubenswrapper[4979]: I1204 13:24:34.832418 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f76457df-phmmr" event={"ID":"016a470f-300c-46b0-9241-f73f843de724","Type":"ContainerDied","Data":"f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed"} Dec 04 13:24:35 crc kubenswrapper[4979]: I1204 13:24:35.029765 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-tcn4f"] Dec 04 13:24:35 crc kubenswrapper[4979]: I1204 13:24:35.039431 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-tcn4f"] Dec 04 13:24:35 crc kubenswrapper[4979]: I1204 13:24:35.931523 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7fd6595c6b-wn2bx" Dec 04 13:24:36 crc kubenswrapper[4979]: I1204 13:24:36.214215 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0522ef0b-76e7-4c30-af05-d7f3f4eb3154" path="/var/lib/kubelet/pods/0522ef0b-76e7-4c30-af05-d7f3f4eb3154/volumes" Dec 04 13:24:44 crc kubenswrapper[4979]: I1204 13:24:44.801756 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f76457df-phmmr" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.902026 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs"] Dec 04 13:24:49 crc kubenswrapper[4979]: E1204 13:24:49.903093 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.903110 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" Dec 04 13:24:49 crc kubenswrapper[4979]: E1204 13:24:49.903154 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon-log" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.903165 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon-log" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.903410 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.903441 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="58966164-dd31-4e68-942e-97f6e9b78574" containerName="horizon-log" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.905154 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.912297 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 04 13:24:49 crc kubenswrapper[4979]: I1204 13:24:49.915201 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs"] Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.012018 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.012522 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdsxq\" (UniqueName: \"kubernetes.io/projected/89020d69-4269-45d7-a016-59f4f74e53a6-kube-api-access-xdsxq\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.012580 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.113992 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.114154 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdsxq\" (UniqueName: \"kubernetes.io/projected/89020d69-4269-45d7-a016-59f4f74e53a6-kube-api-access-xdsxq\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.114175 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.114584 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.114623 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.135058 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdsxq\" (UniqueName: \"kubernetes.io/projected/89020d69-4269-45d7-a016-59f4f74e53a6-kube-api-access-xdsxq\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.232484 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.689950 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs"] Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.989018 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" event={"ID":"89020d69-4269-45d7-a016-59f4f74e53a6","Type":"ContainerStarted","Data":"10751c76769c463a19ec381ce6d1acae265cacffb8686251861c6753d07b7490"} Dec 04 13:24:50 crc kubenswrapper[4979]: I1204 13:24:50.989466 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" event={"ID":"89020d69-4269-45d7-a016-59f4f74e53a6","Type":"ContainerStarted","Data":"941e2140ca299d8c1262cb6a545ecb7ecc0a0478716482e8fd317813d22fb703"} Dec 04 13:24:51 crc kubenswrapper[4979]: I1204 13:24:51.998875 4979 generic.go:334] "Generic (PLEG): container finished" podID="89020d69-4269-45d7-a016-59f4f74e53a6" containerID="10751c76769c463a19ec381ce6d1acae265cacffb8686251861c6753d07b7490" exitCode=0 Dec 04 13:24:51 crc kubenswrapper[4979]: I1204 13:24:51.998938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" event={"ID":"89020d69-4269-45d7-a016-59f4f74e53a6","Type":"ContainerDied","Data":"10751c76769c463a19ec381ce6d1acae265cacffb8686251861c6753d07b7490"} Dec 04 13:24:54 crc kubenswrapper[4979]: I1204 13:24:54.021127 4979 generic.go:334] "Generic (PLEG): container finished" podID="89020d69-4269-45d7-a016-59f4f74e53a6" containerID="49e91c260a19788efd03058f6bacfd25811a9d6e3268a6a68520651a82037abe" exitCode=0 Dec 04 13:24:54 crc kubenswrapper[4979]: I1204 13:24:54.021377 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" event={"ID":"89020d69-4269-45d7-a016-59f4f74e53a6","Type":"ContainerDied","Data":"49e91c260a19788efd03058f6bacfd25811a9d6e3268a6a68520651a82037abe"} Dec 04 13:24:54 crc kubenswrapper[4979]: I1204 13:24:54.801201 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-64f76457df-phmmr" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" probeResult="failure" output="Get \"http://10.217.1.116:8080/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.116:8080: connect: connection refused" Dec 04 13:24:54 crc kubenswrapper[4979]: I1204 13:24:54.801554 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:24:55 crc kubenswrapper[4979]: I1204 13:24:55.036887 4979 generic.go:334] "Generic (PLEG): container finished" podID="89020d69-4269-45d7-a016-59f4f74e53a6" containerID="adf60cea558c2e07fd07bc6d062dc0075bb3dc87ec1b6ee16c213bb6014f583a" exitCode=0 Dec 04 13:24:55 crc kubenswrapper[4979]: I1204 13:24:55.037101 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" event={"ID":"89020d69-4269-45d7-a016-59f4f74e53a6","Type":"ContainerDied","Data":"adf60cea558c2e07fd07bc6d062dc0075bb3dc87ec1b6ee16c213bb6014f583a"} Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.407522 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.544206 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-bundle\") pod \"89020d69-4269-45d7-a016-59f4f74e53a6\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.544614 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-util\") pod \"89020d69-4269-45d7-a016-59f4f74e53a6\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.544697 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdsxq\" (UniqueName: \"kubernetes.io/projected/89020d69-4269-45d7-a016-59f4f74e53a6-kube-api-access-xdsxq\") pod \"89020d69-4269-45d7-a016-59f4f74e53a6\" (UID: \"89020d69-4269-45d7-a016-59f4f74e53a6\") " Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.546406 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-bundle" (OuterVolumeSpecName: "bundle") pod "89020d69-4269-45d7-a016-59f4f74e53a6" (UID: "89020d69-4269-45d7-a016-59f4f74e53a6"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.550012 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89020d69-4269-45d7-a016-59f4f74e53a6-kube-api-access-xdsxq" (OuterVolumeSpecName: "kube-api-access-xdsxq") pod "89020d69-4269-45d7-a016-59f4f74e53a6" (UID: "89020d69-4269-45d7-a016-59f4f74e53a6"). InnerVolumeSpecName "kube-api-access-xdsxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.552533 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-util" (OuterVolumeSpecName: "util") pod "89020d69-4269-45d7-a016-59f4f74e53a6" (UID: "89020d69-4269-45d7-a016-59f4f74e53a6"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.647024 4979 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-util\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.647067 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdsxq\" (UniqueName: \"kubernetes.io/projected/89020d69-4269-45d7-a016-59f4f74e53a6-kube-api-access-xdsxq\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:56 crc kubenswrapper[4979]: I1204 13:24:56.647081 4979 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/89020d69-4269-45d7-a016-59f4f74e53a6-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:24:57 crc kubenswrapper[4979]: I1204 13:24:57.057730 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" event={"ID":"89020d69-4269-45d7-a016-59f4f74e53a6","Type":"ContainerDied","Data":"941e2140ca299d8c1262cb6a545ecb7ecc0a0478716482e8fd317813d22fb703"} Dec 04 13:24:57 crc kubenswrapper[4979]: I1204 13:24:57.057770 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="941e2140ca299d8c1262cb6a545ecb7ecc0a0478716482e8fd317813d22fb703" Dec 04 13:24:57 crc kubenswrapper[4979]: I1204 13:24:57.057806 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs" Dec 04 13:24:58 crc kubenswrapper[4979]: I1204 13:24:58.040442 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:24:58 crc kubenswrapper[4979]: I1204 13:24:58.040778 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.020059 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.095180 4979 generic.go:334] "Generic (PLEG): container finished" podID="016a470f-300c-46b0-9241-f73f843de724" containerID="19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8" exitCode=137 Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.095234 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f76457df-phmmr" event={"ID":"016a470f-300c-46b0-9241-f73f843de724","Type":"ContainerDied","Data":"19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8"} Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.095595 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-64f76457df-phmmr" event={"ID":"016a470f-300c-46b0-9241-f73f843de724","Type":"ContainerDied","Data":"132f429ae86e6a407aaedba6d1e26a0d92736155fa2157b0859c399ae23ff61e"} Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.095625 4979 scope.go:117] "RemoveContainer" containerID="f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.095261 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-64f76457df-phmmr" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.138226 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016a470f-300c-46b0-9241-f73f843de724-logs\") pod \"016a470f-300c-46b0-9241-f73f843de724\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.138604 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/016a470f-300c-46b0-9241-f73f843de724-logs" (OuterVolumeSpecName: "logs") pod "016a470f-300c-46b0-9241-f73f843de724" (UID: "016a470f-300c-46b0-9241-f73f843de724"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.138802 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/016a470f-300c-46b0-9241-f73f843de724-horizon-secret-key\") pod \"016a470f-300c-46b0-9241-f73f843de724\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.138953 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-scripts\") pod \"016a470f-300c-46b0-9241-f73f843de724\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.139066 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-config-data\") pod \"016a470f-300c-46b0-9241-f73f843de724\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.139154 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6rm6\" (UniqueName: \"kubernetes.io/projected/016a470f-300c-46b0-9241-f73f843de724-kube-api-access-g6rm6\") pod \"016a470f-300c-46b0-9241-f73f843de724\" (UID: \"016a470f-300c-46b0-9241-f73f843de724\") " Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.139881 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/016a470f-300c-46b0-9241-f73f843de724-logs\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.144245 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/016a470f-300c-46b0-9241-f73f843de724-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "016a470f-300c-46b0-9241-f73f843de724" (UID: "016a470f-300c-46b0-9241-f73f843de724"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.144512 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/016a470f-300c-46b0-9241-f73f843de724-kube-api-access-g6rm6" (OuterVolumeSpecName: "kube-api-access-g6rm6") pod "016a470f-300c-46b0-9241-f73f843de724" (UID: "016a470f-300c-46b0-9241-f73f843de724"). InnerVolumeSpecName "kube-api-access-g6rm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.165620 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-config-data" (OuterVolumeSpecName: "config-data") pod "016a470f-300c-46b0-9241-f73f843de724" (UID: "016a470f-300c-46b0-9241-f73f843de724"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.166411 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-scripts" (OuterVolumeSpecName: "scripts") pod "016a470f-300c-46b0-9241-f73f843de724" (UID: "016a470f-300c-46b0-9241-f73f843de724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.241485 4979 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/016a470f-300c-46b0-9241-f73f843de724-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.241521 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.241529 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/016a470f-300c-46b0-9241-f73f843de724-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.241541 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6rm6\" (UniqueName: \"kubernetes.io/projected/016a470f-300c-46b0-9241-f73f843de724-kube-api-access-g6rm6\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.280336 4979 scope.go:117] "RemoveContainer" containerID="19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.300739 4979 scope.go:117] "RemoveContainer" containerID="f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed" Dec 04 13:25:01 crc kubenswrapper[4979]: E1204 13:25:01.302117 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed\": container with ID starting with f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed not found: ID does not exist" containerID="f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.302186 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed"} err="failed to get container status \"f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed\": rpc error: code = NotFound desc = could not find container \"f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed\": container with ID starting with f5ee03e36cddbe49c7ca17795ad0c739179804f42ee470acff5b146835a893ed not found: ID does not exist" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.302213 4979 scope.go:117] "RemoveContainer" containerID="19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8" Dec 04 13:25:01 crc kubenswrapper[4979]: E1204 13:25:01.302641 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8\": container with ID starting with 19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8 not found: ID does not exist" containerID="19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.302687 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8"} err="failed to get container status \"19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8\": rpc error: code = NotFound desc = could not find container \"19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8\": container with ID starting with 19eb2c12f8229734c5d49997f6ce552df18171ed964ac745b8fb6f6acb7001f8 not found: ID does not exist" Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.448798 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-64f76457df-phmmr"] Dec 04 13:25:01 crc kubenswrapper[4979]: I1204 13:25:01.749389 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-64f76457df-phmmr"] Dec 04 13:25:02 crc kubenswrapper[4979]: I1204 13:25:02.212503 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="016a470f-300c-46b0-9241-f73f843de724" path="/var/lib/kubelet/pods/016a470f-300c-46b0-9241-f73f843de724/volumes" Dec 04 13:25:05 crc kubenswrapper[4979]: I1204 13:25:05.065916 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-2d60-account-create-update-5vtf9"] Dec 04 13:25:05 crc kubenswrapper[4979]: I1204 13:25:05.089186 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-7kpc6"] Dec 04 13:25:05 crc kubenswrapper[4979]: I1204 13:25:05.099048 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-2d60-account-create-update-5vtf9"] Dec 04 13:25:05 crc kubenswrapper[4979]: I1204 13:25:05.110078 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-7kpc6"] Dec 04 13:25:06 crc kubenswrapper[4979]: I1204 13:25:06.211312 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6f58092-8c55-4172-b2b5-da79f085277c" path="/var/lib/kubelet/pods/a6f58092-8c55-4172-b2b5-da79f085277c/volumes" Dec 04 13:25:06 crc kubenswrapper[4979]: I1204 13:25:06.213360 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7043a66-6d3c-423a-a159-84afd283492e" path="/var/lib/kubelet/pods/e7043a66-6d3c-423a-a159-84afd283492e/volumes" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.309592 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5"] Dec 04 13:25:08 crc kubenswrapper[4979]: E1204 13:25:08.310243 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="util" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310255 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="util" Dec 04 13:25:08 crc kubenswrapper[4979]: E1204 13:25:08.310280 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="pull" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310286 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="pull" Dec 04 13:25:08 crc kubenswrapper[4979]: E1204 13:25:08.310319 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310328 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" Dec 04 13:25:08 crc kubenswrapper[4979]: E1204 13:25:08.310342 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="extract" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310348 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="extract" Dec 04 13:25:08 crc kubenswrapper[4979]: E1204 13:25:08.310362 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon-log" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310367 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon-log" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310537 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon-log" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310556 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="016a470f-300c-46b0-9241-f73f843de724" containerName="horizon" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.310565 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="89020d69-4269-45d7-a016-59f4f74e53a6" containerName="extract" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.311193 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.317656 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.317706 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.317829 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-jpdjx" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.332769 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.407442 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j522n\" (UniqueName: \"kubernetes.io/projected/e2d039ea-6c89-4435-88ea-e5ef847fe97c-kube-api-access-j522n\") pod \"obo-prometheus-operator-668cf9dfbb-5wrx5\" (UID: \"e2d039ea-6c89-4435-88ea-e5ef847fe97c\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.486104 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.487702 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.491733 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.493652 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-l62mg" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.502453 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.504117 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.508765 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j522n\" (UniqueName: \"kubernetes.io/projected/e2d039ea-6c89-4435-88ea-e5ef847fe97c-kube-api-access-j522n\") pod \"obo-prometheus-operator-668cf9dfbb-5wrx5\" (UID: \"e2d039ea-6c89-4435-88ea-e5ef847fe97c\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.511156 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.556762 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j522n\" (UniqueName: \"kubernetes.io/projected/e2d039ea-6c89-4435-88ea-e5ef847fe97c-kube-api-access-j522n\") pod \"obo-prometheus-operator-668cf9dfbb-5wrx5\" (UID: \"e2d039ea-6c89-4435-88ea-e5ef847fe97c\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.562859 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.610908 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0e7abc19-0af1-4edc-89bd-4ec0ab996bac-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh\" (UID: \"0e7abc19-0af1-4edc-89bd-4ec0ab996bac\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.611020 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0e7abc19-0af1-4edc-89bd-4ec0ab996bac-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh\" (UID: \"0e7abc19-0af1-4edc-89bd-4ec0ab996bac\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.611102 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a3ef5ba-d73a-42e2-9942-6ba8fc02da27-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv\" (UID: \"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.611180 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a3ef5ba-d73a-42e2-9942-6ba8fc02da27-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv\" (UID: \"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.635783 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.649055 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-clrnn"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.650529 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.657445 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.657667 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-g8mvt" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.666622 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-clrnn"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.713248 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a3ef5ba-d73a-42e2-9942-6ba8fc02da27-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv\" (UID: \"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.713648 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0e7abc19-0af1-4edc-89bd-4ec0ab996bac-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh\" (UID: \"0e7abc19-0af1-4edc-89bd-4ec0ab996bac\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.713701 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0e7abc19-0af1-4edc-89bd-4ec0ab996bac-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh\" (UID: \"0e7abc19-0af1-4edc-89bd-4ec0ab996bac\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.713758 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a3ef5ba-d73a-42e2-9942-6ba8fc02da27-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv\" (UID: \"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.718692 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/8a3ef5ba-d73a-42e2-9942-6ba8fc02da27-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv\" (UID: \"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.718857 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0e7abc19-0af1-4edc-89bd-4ec0ab996bac-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh\" (UID: \"0e7abc19-0af1-4edc-89bd-4ec0ab996bac\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.719733 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/8a3ef5ba-d73a-42e2-9942-6ba8fc02da27-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv\" (UID: \"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.719992 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0e7abc19-0af1-4edc-89bd-4ec0ab996bac-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh\" (UID: \"0e7abc19-0af1-4edc-89bd-4ec0ab996bac\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.807219 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.816689 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0264a7ca-8bc9-43e6-8995-885f46a759dc-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-clrnn\" (UID: \"0264a7ca-8bc9-43e6-8995-885f46a759dc\") " pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.817205 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5m52\" (UniqueName: \"kubernetes.io/projected/0264a7ca-8bc9-43e6-8995-885f46a759dc-kube-api-access-b5m52\") pod \"observability-operator-d8bb48f5d-clrnn\" (UID: \"0264a7ca-8bc9-43e6-8995-885f46a759dc\") " pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.826849 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.874483 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-n6rcx"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.875940 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.880049 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-lxgtp" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.880933 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-n6rcx"] Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.919134 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0264a7ca-8bc9-43e6-8995-885f46a759dc-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-clrnn\" (UID: \"0264a7ca-8bc9-43e6-8995-885f46a759dc\") " pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.919663 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5m52\" (UniqueName: \"kubernetes.io/projected/0264a7ca-8bc9-43e6-8995-885f46a759dc-kube-api-access-b5m52\") pod \"observability-operator-d8bb48f5d-clrnn\" (UID: \"0264a7ca-8bc9-43e6-8995-885f46a759dc\") " pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.928811 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/0264a7ca-8bc9-43e6-8995-885f46a759dc-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-clrnn\" (UID: \"0264a7ca-8bc9-43e6-8995-885f46a759dc\") " pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:08 crc kubenswrapper[4979]: I1204 13:25:08.942452 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5m52\" (UniqueName: \"kubernetes.io/projected/0264a7ca-8bc9-43e6-8995-885f46a759dc-kube-api-access-b5m52\") pod \"observability-operator-d8bb48f5d-clrnn\" (UID: \"0264a7ca-8bc9-43e6-8995-885f46a759dc\") " pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.021733 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/eeec64e3-82f9-4370-8621-fa7963371a38-openshift-service-ca\") pod \"perses-operator-5446b9c989-n6rcx\" (UID: \"eeec64e3-82f9-4370-8621-fa7963371a38\") " pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.021894 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rcts\" (UniqueName: \"kubernetes.io/projected/eeec64e3-82f9-4370-8621-fa7963371a38-kube-api-access-9rcts\") pod \"perses-operator-5446b9c989-n6rcx\" (UID: \"eeec64e3-82f9-4370-8621-fa7963371a38\") " pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.125040 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rcts\" (UniqueName: \"kubernetes.io/projected/eeec64e3-82f9-4370-8621-fa7963371a38-kube-api-access-9rcts\") pod \"perses-operator-5446b9c989-n6rcx\" (UID: \"eeec64e3-82f9-4370-8621-fa7963371a38\") " pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.125135 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/eeec64e3-82f9-4370-8621-fa7963371a38-openshift-service-ca\") pod \"perses-operator-5446b9c989-n6rcx\" (UID: \"eeec64e3-82f9-4370-8621-fa7963371a38\") " pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.125998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/eeec64e3-82f9-4370-8621-fa7963371a38-openshift-service-ca\") pod \"perses-operator-5446b9c989-n6rcx\" (UID: \"eeec64e3-82f9-4370-8621-fa7963371a38\") " pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.147217 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rcts\" (UniqueName: \"kubernetes.io/projected/eeec64e3-82f9-4370-8621-fa7963371a38-kube-api-access-9rcts\") pod \"perses-operator-5446b9c989-n6rcx\" (UID: \"eeec64e3-82f9-4370-8621-fa7963371a38\") " pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.154559 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.178785 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5"] Dec 04 13:25:09 crc kubenswrapper[4979]: W1204 13:25:09.199611 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2d039ea_6c89_4435_88ea_e5ef847fe97c.slice/crio-af8945c9ec9df171c12bb9bab19449eae1ce80f6f5f288d28ddd414b2fbbfae5 WatchSource:0}: Error finding container af8945c9ec9df171c12bb9bab19449eae1ce80f6f5f288d28ddd414b2fbbfae5: Status 404 returned error can't find the container with id af8945c9ec9df171c12bb9bab19449eae1ce80f6f5f288d28ddd414b2fbbfae5 Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.278812 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.664536 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh"] Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.966537 4979 scope.go:117] "RemoveContainer" containerID="fa195f06e4e43bbbbb877669aa5457baf431b7f3a779e0f1cc1d35e2d85c6388" Dec 04 13:25:09 crc kubenswrapper[4979]: I1204 13:25:09.988967 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv"] Dec 04 13:25:09 crc kubenswrapper[4979]: W1204 13:25:09.992523 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a3ef5ba_d73a_42e2_9942_6ba8fc02da27.slice/crio-8dfd0095e7b4f654d9d3569166fe9692d9bd57ed7d3b9a0e00a4ea08361d7c8c WatchSource:0}: Error finding container 8dfd0095e7b4f654d9d3569166fe9692d9bd57ed7d3b9a0e00a4ea08361d7c8c: Status 404 returned error can't find the container with id 8dfd0095e7b4f654d9d3569166fe9692d9bd57ed7d3b9a0e00a4ea08361d7c8c Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.013564 4979 scope.go:117] "RemoveContainer" containerID="2eba55fc70887aaf607b256f7a7f1a7939db931f489376c840e3549523186743" Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.131983 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-n6rcx"] Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.143174 4979 scope.go:117] "RemoveContainer" containerID="28c2582ba35345340d3fbe83f285b843289af55b4163fcc7e49865be02c4a8ee" Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.151782 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-clrnn"] Dec 04 13:25:10 crc kubenswrapper[4979]: W1204 13:25:10.155363 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeeec64e3_82f9_4370_8621_fa7963371a38.slice/crio-f69cd2713c89a18f283924e2fe94458d4abe69e42ea5fd67c7e66a2b9d39a0f7 WatchSource:0}: Error finding container f69cd2713c89a18f283924e2fe94458d4abe69e42ea5fd67c7e66a2b9d39a0f7: Status 404 returned error can't find the container with id f69cd2713c89a18f283924e2fe94458d4abe69e42ea5fd67c7e66a2b9d39a0f7 Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.193391 4979 scope.go:117] "RemoveContainer" containerID="a82cef0dea9e1aa128f51638af2ab6923edf92bdea5fbfb84c6fd95c6e0d4ae0" Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.230032 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" event={"ID":"0e7abc19-0af1-4edc-89bd-4ec0ab996bac","Type":"ContainerStarted","Data":"48f14bf3bbc1459f8a7b700a6946e8733fb82647aef5d4ddbcee493d30bfef2b"} Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.230090 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" event={"ID":"eeec64e3-82f9-4370-8621-fa7963371a38","Type":"ContainerStarted","Data":"f69cd2713c89a18f283924e2fe94458d4abe69e42ea5fd67c7e66a2b9d39a0f7"} Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.253541 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" event={"ID":"0264a7ca-8bc9-43e6-8995-885f46a759dc","Type":"ContainerStarted","Data":"19ca6213279451c15f151927fb15238656d75f175508477f0a5bf3bea42a4f80"} Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.254416 4979 scope.go:117] "RemoveContainer" containerID="af27934eeef91f7c26f220e51ad6e7a0708a3f9d6c356153b2cf2eb76c747a1e" Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.273620 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" event={"ID":"e2d039ea-6c89-4435-88ea-e5ef847fe97c","Type":"ContainerStarted","Data":"af8945c9ec9df171c12bb9bab19449eae1ce80f6f5f288d28ddd414b2fbbfae5"} Dec 04 13:25:10 crc kubenswrapper[4979]: I1204 13:25:10.275926 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" event={"ID":"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27","Type":"ContainerStarted","Data":"8dfd0095e7b4f654d9d3569166fe9692d9bd57ed7d3b9a0e00a4ea08361d7c8c"} Dec 04 13:25:12 crc kubenswrapper[4979]: I1204 13:25:12.055057 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-f627f"] Dec 04 13:25:12 crc kubenswrapper[4979]: I1204 13:25:12.070723 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-f627f"] Dec 04 13:25:12 crc kubenswrapper[4979]: I1204 13:25:12.230778 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8c7039b-417d-4634-b1ca-8a145614d908" path="/var/lib/kubelet/pods/d8c7039b-417d-4634-b1ca-8a145614d908/volumes" Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.364051 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" event={"ID":"eeec64e3-82f9-4370-8621-fa7963371a38","Type":"ContainerStarted","Data":"9bb86a2f0a0eb0667c81d682c725a50b2190d835dca643ce99e51cbdaf63c183"} Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.364764 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.371121 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" event={"ID":"e2d039ea-6c89-4435-88ea-e5ef847fe97c","Type":"ContainerStarted","Data":"f1e8d3480575fe04c20723d741601585ec05233a948c1373397a982615b70cd8"} Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.384728 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" event={"ID":"8a3ef5ba-d73a-42e2-9942-6ba8fc02da27","Type":"ContainerStarted","Data":"3ce0aebcd61445dbd15d639fc699e724e7606f219fbd9c4663deb30bf694b821"} Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.386206 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" podStartSLOduration=3.240626035 podStartE2EDuration="8.38619146s" podCreationTimestamp="2025-12-04 13:25:08 +0000 UTC" firstStartedPulling="2025-12-04 13:25:10.171385424 +0000 UTC m=+6134.445681228" lastFinishedPulling="2025-12-04 13:25:15.316950849 +0000 UTC m=+6139.591246653" observedRunningTime="2025-12-04 13:25:16.381914686 +0000 UTC m=+6140.656210500" watchObservedRunningTime="2025-12-04 13:25:16.38619146 +0000 UTC m=+6140.660487264" Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.391765 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" event={"ID":"0e7abc19-0af1-4edc-89bd-4ec0ab996bac","Type":"ContainerStarted","Data":"c798c8cf9b84b032aa823abe06440b09d03660db262ad5ef55853c98b209aa82"} Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.416114 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-5wrx5" podStartSLOduration=2.317234145 podStartE2EDuration="8.416095542s" podCreationTimestamp="2025-12-04 13:25:08 +0000 UTC" firstStartedPulling="2025-12-04 13:25:09.218963596 +0000 UTC m=+6133.493259400" lastFinishedPulling="2025-12-04 13:25:15.317824993 +0000 UTC m=+6139.592120797" observedRunningTime="2025-12-04 13:25:16.398680525 +0000 UTC m=+6140.672976349" watchObservedRunningTime="2025-12-04 13:25:16.416095542 +0000 UTC m=+6140.690391346" Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.442874 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv" podStartSLOduration=3.160089325 podStartE2EDuration="8.442852059s" podCreationTimestamp="2025-12-04 13:25:08 +0000 UTC" firstStartedPulling="2025-12-04 13:25:10.013516441 +0000 UTC m=+6134.287812245" lastFinishedPulling="2025-12-04 13:25:15.296279175 +0000 UTC m=+6139.570574979" observedRunningTime="2025-12-04 13:25:16.421708572 +0000 UTC m=+6140.696004376" watchObservedRunningTime="2025-12-04 13:25:16.442852059 +0000 UTC m=+6140.717147863" Dec 04 13:25:16 crc kubenswrapper[4979]: I1204 13:25:16.452042 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh" podStartSLOduration=2.88131986 podStartE2EDuration="8.452016155s" podCreationTimestamp="2025-12-04 13:25:08 +0000 UTC" firstStartedPulling="2025-12-04 13:25:09.695912064 +0000 UTC m=+6133.970207868" lastFinishedPulling="2025-12-04 13:25:15.266608349 +0000 UTC m=+6139.540904163" observedRunningTime="2025-12-04 13:25:16.446714123 +0000 UTC m=+6140.721009917" watchObservedRunningTime="2025-12-04 13:25:16.452016155 +0000 UTC m=+6140.726311959" Dec 04 13:25:19 crc kubenswrapper[4979]: I1204 13:25:19.422447 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" event={"ID":"0264a7ca-8bc9-43e6-8995-885f46a759dc","Type":"ContainerStarted","Data":"07ba894b80e019121780a5e68d16572eb572c3f792f2afe035ce95bd4623b701"} Dec 04 13:25:19 crc kubenswrapper[4979]: I1204 13:25:19.423068 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:19 crc kubenswrapper[4979]: I1204 13:25:19.425342 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" Dec 04 13:25:19 crc kubenswrapper[4979]: I1204 13:25:19.450703 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-clrnn" podStartSLOduration=3.146871551 podStartE2EDuration="11.450678022s" podCreationTimestamp="2025-12-04 13:25:08 +0000 UTC" firstStartedPulling="2025-12-04 13:25:10.173572723 +0000 UTC m=+6134.447868527" lastFinishedPulling="2025-12-04 13:25:18.477379194 +0000 UTC m=+6142.751674998" observedRunningTime="2025-12-04 13:25:19.447267901 +0000 UTC m=+6143.721563705" watchObservedRunningTime="2025-12-04 13:25:19.450678022 +0000 UTC m=+6143.724973866" Dec 04 13:25:28 crc kubenswrapper[4979]: I1204 13:25:28.040884 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:25:28 crc kubenswrapper[4979]: I1204 13:25:28.041445 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:25:29 crc kubenswrapper[4979]: I1204 13:25:29.282844 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.798928 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.799534 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstackclient" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" containerName="openstackclient" containerID="cri-o://1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac" gracePeriod=2 Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.812231 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.841945 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 04 13:25:31 crc kubenswrapper[4979]: E1204 13:25:31.842491 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" containerName="openstackclient" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.842537 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" containerName="openstackclient" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.842785 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" containerName="openstackclient" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.843672 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.848716 4979 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="380f708e-bc42-431f-b675-af01d9c11ae8" podUID="e55a2ea6-d990-412d-b5fb-21631921a849" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.858634 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.962518 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e55a2ea6-d990-412d-b5fb-21631921a849-openstack-config\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.963038 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e55a2ea6-d990-412d-b5fb-21631921a849-openstack-config-secret\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:31 crc kubenswrapper[4979]: I1204 13:25:31.963117 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vf5r4\" (UniqueName: \"kubernetes.io/projected/e55a2ea6-d990-412d-b5fb-21631921a849-kube-api-access-vf5r4\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.048842 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.050374 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.052986 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-c2phj" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.064950 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e55a2ea6-d990-412d-b5fb-21631921a849-openstack-config\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.065049 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e55a2ea6-d990-412d-b5fb-21631921a849-openstack-config-secret\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.065087 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vf5r4\" (UniqueName: \"kubernetes.io/projected/e55a2ea6-d990-412d-b5fb-21631921a849-kube-api-access-vf5r4\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.066203 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/e55a2ea6-d990-412d-b5fb-21631921a849-openstack-config\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.068883 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.073999 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/e55a2ea6-d990-412d-b5fb-21631921a849-openstack-config-secret\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.086054 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vf5r4\" (UniqueName: \"kubernetes.io/projected/e55a2ea6-d990-412d-b5fb-21631921a849-kube-api-access-vf5r4\") pod \"openstackclient\" (UID: \"e55a2ea6-d990-412d-b5fb-21631921a849\") " pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.166538 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn7ps\" (UniqueName: \"kubernetes.io/projected/116e4048-6b4c-42b7-a6d8-f0b436b1b176-kube-api-access-bn7ps\") pod \"kube-state-metrics-0\" (UID: \"116e4048-6b4c-42b7-a6d8-f0b436b1b176\") " pod="openstack/kube-state-metrics-0" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.170108 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.270909 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn7ps\" (UniqueName: \"kubernetes.io/projected/116e4048-6b4c-42b7-a6d8-f0b436b1b176-kube-api-access-bn7ps\") pod \"kube-state-metrics-0\" (UID: \"116e4048-6b4c-42b7-a6d8-f0b436b1b176\") " pod="openstack/kube-state-metrics-0" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.316291 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn7ps\" (UniqueName: \"kubernetes.io/projected/116e4048-6b4c-42b7-a6d8-f0b436b1b176-kube-api-access-bn7ps\") pod \"kube-state-metrics-0\" (UID: \"116e4048-6b4c-42b7-a6d8-f0b436b1b176\") " pod="openstack/kube-state-metrics-0" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.371785 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.990371 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 04 13:25:32 crc kubenswrapper[4979]: I1204 13:25:32.995787 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.023372 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-cluster-tls-config" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.023437 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-generated" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.023483 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-alertmanager-dockercfg-25cvr" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.023506 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-web-config" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.023564 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"alertmanager-metric-storage-tls-assets-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.039181 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.065709 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126565 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126628 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126670 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cb54v\" (UniqueName: \"kubernetes.io/projected/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-kube-api-access-cb54v\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126717 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126757 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126772 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.126845 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230486 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230540 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230564 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230635 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230668 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230694 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.230726 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cb54v\" (UniqueName: \"kubernetes.io/projected/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-kube-api-access-cb54v\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.231479 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-metric-storage-db\" (UniqueName: \"kubernetes.io/empty-dir/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-alertmanager-metric-storage-db\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.240123 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-config-volume\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.248016 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-tls-assets\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.254483 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-web-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.263961 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-config-out\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.264798 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-cluster-tls-config\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.312887 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cb54v\" (UniqueName: \"kubernetes.io/projected/e9693784-a0d4-4fa6-8941-fd9f6617ae6c-kube-api-access-cb54v\") pod \"alertmanager-metric-storage-0\" (UID: \"e9693784-a0d4-4fa6-8941-fd9f6617ae6c\") " pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.415927 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/alertmanager-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.499247 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.502505 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.510102 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.510423 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.510634 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.520176 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.520447 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-9s4ts" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.544940 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.580739 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.622405 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.637404 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e55a2ea6-d990-412d-b5fb-21631921a849","Type":"ContainerStarted","Data":"b596c4a72648769f2c544bb2af8cb4aa9a53bfcf26d51fe0a2f646b989ee6540"} Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.647964 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/116becdc-fe97-4de3-902f-77dfd1332c5b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648058 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvsft\" (UniqueName: \"kubernetes.io/projected/116becdc-fe97-4de3-902f-77dfd1332c5b-kube-api-access-vvsft\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648113 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648140 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648182 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/116becdc-fe97-4de3-902f-77dfd1332c5b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648380 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648515 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/116becdc-fe97-4de3-902f-77dfd1332c5b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.648537 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-config\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750328 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/116becdc-fe97-4de3-902f-77dfd1332c5b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750386 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-config\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750485 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/116becdc-fe97-4de3-902f-77dfd1332c5b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750556 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvsft\" (UniqueName: \"kubernetes.io/projected/116becdc-fe97-4de3-902f-77dfd1332c5b-kube-api-access-vvsft\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750597 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750628 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750656 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/116becdc-fe97-4de3-902f-77dfd1332c5b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.750704 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.754683 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/116becdc-fe97-4de3-902f-77dfd1332c5b-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.761953 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/116becdc-fe97-4de3-902f-77dfd1332c5b-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.763738 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.764083 4979 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.764147 4979 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/c1ee614d1d1587f2f73a83cfe8b12cfef3c75cd27a6c6f054f33a94920651d5b/globalmount\"" pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.768886 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-config\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.769284 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/116becdc-fe97-4de3-902f-77dfd1332c5b-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.776795 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/116becdc-fe97-4de3-902f-77dfd1332c5b-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:33 crc kubenswrapper[4979]: I1204 13:25:33.781378 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvsft\" (UniqueName: \"kubernetes.io/projected/116becdc-fe97-4de3-902f-77dfd1332c5b-kube-api-access-vvsft\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.005100 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-bc72b729-3a87-4dec-92e4-5e140baa905f\") pod \"prometheus-metric-storage-0\" (UID: \"116becdc-fe97-4de3-902f-77dfd1332c5b\") " pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.180280 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.317054 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/alertmanager-metric-storage-0"] Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.502058 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.576760 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config-secret\") pod \"380f708e-bc42-431f-b675-af01d9c11ae8\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.577158 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config\") pod \"380f708e-bc42-431f-b675-af01d9c11ae8\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.577211 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpws4\" (UniqueName: \"kubernetes.io/projected/380f708e-bc42-431f-b675-af01d9c11ae8-kube-api-access-hpws4\") pod \"380f708e-bc42-431f-b675-af01d9c11ae8\" (UID: \"380f708e-bc42-431f-b675-af01d9c11ae8\") " Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.590640 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/380f708e-bc42-431f-b675-af01d9c11ae8-kube-api-access-hpws4" (OuterVolumeSpecName: "kube-api-access-hpws4") pod "380f708e-bc42-431f-b675-af01d9c11ae8" (UID: "380f708e-bc42-431f-b675-af01d9c11ae8"). InnerVolumeSpecName "kube-api-access-hpws4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.614248 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "380f708e-bc42-431f-b675-af01d9c11ae8" (UID: "380f708e-bc42-431f-b675-af01d9c11ae8"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.646239 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "380f708e-bc42-431f-b675-af01d9c11ae8" (UID: "380f708e-bc42-431f-b675-af01d9c11ae8"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.651455 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"116e4048-6b4c-42b7-a6d8-f0b436b1b176","Type":"ContainerStarted","Data":"314db85f84d0f6eb23fda17601bcbc9a838962f3aae5ea93fb9910eb2e92256c"} Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.651675 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.651776 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"116e4048-6b4c-42b7-a6d8-f0b436b1b176","Type":"ContainerStarted","Data":"6c4f15b53e73405d4dda270355d7b1c4a05a45ef4dbd6cb8b6344f4f95a14d03"} Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.653246 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"e55a2ea6-d990-412d-b5fb-21631921a849","Type":"ContainerStarted","Data":"6a66b173708314062037d5ae8a901b5c228957fe4474488608e6721c56a8a624"} Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.655018 4979 generic.go:334] "Generic (PLEG): container finished" podID="380f708e-bc42-431f-b675-af01d9c11ae8" containerID="1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac" exitCode=137 Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.655060 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.655097 4979 scope.go:117] "RemoveContainer" containerID="1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.656826 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e9693784-a0d4-4fa6-8941-fd9f6617ae6c","Type":"ContainerStarted","Data":"750a14de5b1b320c567bb6dcd862e3ea7b2b90a001bbc3b02b05ae2d74fc122b"} Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.672060 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.202523975 podStartE2EDuration="2.672038655s" podCreationTimestamp="2025-12-04 13:25:32 +0000 UTC" firstStartedPulling="2025-12-04 13:25:33.640189836 +0000 UTC m=+6157.914485640" lastFinishedPulling="2025-12-04 13:25:34.109704516 +0000 UTC m=+6158.384000320" observedRunningTime="2025-12-04 13:25:34.671052448 +0000 UTC m=+6158.945348252" watchObservedRunningTime="2025-12-04 13:25:34.672038655 +0000 UTC m=+6158.946334459" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.680527 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.682525 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpws4\" (UniqueName: \"kubernetes.io/projected/380f708e-bc42-431f-b675-af01d9c11ae8-kube-api-access-hpws4\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.682607 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/380f708e-bc42-431f-b675-af01d9c11ae8-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.698732 4979 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="380f708e-bc42-431f-b675-af01d9c11ae8" podUID="e55a2ea6-d990-412d-b5fb-21631921a849" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.701240 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.701213546 podStartE2EDuration="3.701213546s" podCreationTimestamp="2025-12-04 13:25:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:25:34.692854042 +0000 UTC m=+6158.967149846" watchObservedRunningTime="2025-12-04 13:25:34.701213546 +0000 UTC m=+6158.975509340" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.705671 4979 scope.go:117] "RemoveContainer" containerID="1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac" Dec 04 13:25:34 crc kubenswrapper[4979]: E1204 13:25:34.707214 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac\": container with ID starting with 1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac not found: ID does not exist" containerID="1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.707326 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac"} err="failed to get container status \"1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac\": rpc error: code = NotFound desc = could not find container \"1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac\": container with ID starting with 1dce6a781ee53c884b89b6cf38a8746dbc6172f65ab885ce2d537b6591d5d0ac not found: ID does not exist" Dec 04 13:25:34 crc kubenswrapper[4979]: I1204 13:25:34.827613 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 04 13:25:34 crc kubenswrapper[4979]: W1204 13:25:34.830521 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod116becdc_fe97_4de3_902f_77dfd1332c5b.slice/crio-da982c1b308c50a1043326952c27927163b5b363f4bc16c937e0794b761d0705 WatchSource:0}: Error finding container da982c1b308c50a1043326952c27927163b5b363f4bc16c937e0794b761d0705: Status 404 returned error can't find the container with id da982c1b308c50a1043326952c27927163b5b363f4bc16c937e0794b761d0705 Dec 04 13:25:35 crc kubenswrapper[4979]: I1204 13:25:35.670400 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"116becdc-fe97-4de3-902f-77dfd1332c5b","Type":"ContainerStarted","Data":"da982c1b308c50a1043326952c27927163b5b363f4bc16c937e0794b761d0705"} Dec 04 13:25:36 crc kubenswrapper[4979]: I1204 13:25:36.209948 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="380f708e-bc42-431f-b675-af01d9c11ae8" path="/var/lib/kubelet/pods/380f708e-bc42-431f-b675-af01d9c11ae8/volumes" Dec 04 13:25:40 crc kubenswrapper[4979]: I1204 13:25:40.735255 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"116becdc-fe97-4de3-902f-77dfd1332c5b","Type":"ContainerStarted","Data":"74dd72bd60f625d219dc6bab2408b8f19bdf04f8f4053d895f413dec4a54a282"} Dec 04 13:25:40 crc kubenswrapper[4979]: I1204 13:25:40.737711 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e9693784-a0d4-4fa6-8941-fd9f6617ae6c","Type":"ContainerStarted","Data":"854d109436d399c8bb1ab38ebe3a545b6982fa4c6404c64907b9bbf461a47a97"} Dec 04 13:25:42 crc kubenswrapper[4979]: I1204 13:25:42.378458 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 04 13:25:46 crc kubenswrapper[4979]: I1204 13:25:46.800706 4979 generic.go:334] "Generic (PLEG): container finished" podID="116becdc-fe97-4de3-902f-77dfd1332c5b" containerID="74dd72bd60f625d219dc6bab2408b8f19bdf04f8f4053d895f413dec4a54a282" exitCode=0 Dec 04 13:25:46 crc kubenswrapper[4979]: I1204 13:25:46.800807 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"116becdc-fe97-4de3-902f-77dfd1332c5b","Type":"ContainerDied","Data":"74dd72bd60f625d219dc6bab2408b8f19bdf04f8f4053d895f413dec4a54a282"} Dec 04 13:25:46 crc kubenswrapper[4979]: I1204 13:25:46.804048 4979 generic.go:334] "Generic (PLEG): container finished" podID="e9693784-a0d4-4fa6-8941-fd9f6617ae6c" containerID="854d109436d399c8bb1ab38ebe3a545b6982fa4c6404c64907b9bbf461a47a97" exitCode=0 Dec 04 13:25:46 crc kubenswrapper[4979]: I1204 13:25:46.804105 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e9693784-a0d4-4fa6-8941-fd9f6617ae6c","Type":"ContainerDied","Data":"854d109436d399c8bb1ab38ebe3a545b6982fa4c6404c64907b9bbf461a47a97"} Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.040798 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.041344 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.041412 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.042227 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"be34e428814c510aac7c63fa5d7dc77256bd5b17bcd712eed8e9b57775e43b6a"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.042288 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://be34e428814c510aac7c63fa5d7dc77256bd5b17bcd712eed8e9b57775e43b6a" gracePeriod=600 Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.917640 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="be34e428814c510aac7c63fa5d7dc77256bd5b17bcd712eed8e9b57775e43b6a" exitCode=0 Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.917701 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"be34e428814c510aac7c63fa5d7dc77256bd5b17bcd712eed8e9b57775e43b6a"} Dec 04 13:25:58 crc kubenswrapper[4979]: I1204 13:25:58.918013 4979 scope.go:117] "RemoveContainer" containerID="9aaac998c512871e12426ebff1d4ab27edf6cdc8ba0d7c4c4b1f0dfb0738f005" Dec 04 13:25:59 crc kubenswrapper[4979]: I1204 13:25:59.929985 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e9693784-a0d4-4fa6-8941-fd9f6617ae6c","Type":"ContainerStarted","Data":"81c81e3f9ed039d32442355706ab2faa2bc23074938acbbc9c1ed4c72b8435c4"} Dec 04 13:25:59 crc kubenswrapper[4979]: I1204 13:25:59.931962 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"116becdc-fe97-4de3-902f-77dfd1332c5b","Type":"ContainerStarted","Data":"9349e6abab256d1a813c2b2fab5bd004463b1bf7ffdc6f1ddec22ac61a0fea5a"} Dec 04 13:25:59 crc kubenswrapper[4979]: I1204 13:25:59.934171 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7"} Dec 04 13:26:02 crc kubenswrapper[4979]: I1204 13:26:02.989477 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"116becdc-fe97-4de3-902f-77dfd1332c5b","Type":"ContainerStarted","Data":"74ca2ea43573ba46ab853a7f00ba7731207daf12e63ab01be86f6735c534b846"} Dec 04 13:26:02 crc kubenswrapper[4979]: I1204 13:26:02.993458 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/alertmanager-metric-storage-0" event={"ID":"e9693784-a0d4-4fa6-8941-fd9f6617ae6c","Type":"ContainerStarted","Data":"daa792e4f9e128a8576e5e40d14d559b4a112841c4b182a72578dfbf3b69d8c2"} Dec 04 13:26:02 crc kubenswrapper[4979]: I1204 13:26:02.994024 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/alertmanager-metric-storage-0" Dec 04 13:26:02 crc kubenswrapper[4979]: I1204 13:26:02.997390 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/alertmanager-metric-storage-0" Dec 04 13:26:03 crc kubenswrapper[4979]: I1204 13:26:03.024667 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/alertmanager-metric-storage-0" podStartSLOduration=6.583797395 podStartE2EDuration="31.024649892s" podCreationTimestamp="2025-12-04 13:25:32 +0000 UTC" firstStartedPulling="2025-12-04 13:25:34.364536669 +0000 UTC m=+6158.638832473" lastFinishedPulling="2025-12-04 13:25:58.805389166 +0000 UTC m=+6183.079684970" observedRunningTime="2025-12-04 13:26:03.01860677 +0000 UTC m=+6187.292902574" watchObservedRunningTime="2025-12-04 13:26:03.024649892 +0000 UTC m=+6187.298945696" Dec 04 13:26:06 crc kubenswrapper[4979]: I1204 13:26:06.027804 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"116becdc-fe97-4de3-902f-77dfd1332c5b","Type":"ContainerStarted","Data":"5b793890babb7091a949af0c3df0cfc458652ad1d437b737c36666916898a6dc"} Dec 04 13:26:06 crc kubenswrapper[4979]: I1204 13:26:06.063387 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=3.977432778 podStartE2EDuration="34.063366014s" podCreationTimestamp="2025-12-04 13:25:32 +0000 UTC" firstStartedPulling="2025-12-04 13:25:34.83263863 +0000 UTC m=+6159.106934434" lastFinishedPulling="2025-12-04 13:26:04.918571846 +0000 UTC m=+6189.192867670" observedRunningTime="2025-12-04 13:26:06.057868776 +0000 UTC m=+6190.332164580" watchObservedRunningTime="2025-12-04 13:26:06.063366014 +0000 UTC m=+6190.337661818" Dec 04 13:26:09 crc kubenswrapper[4979]: I1204 13:26:09.181991 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 04 13:26:10 crc kubenswrapper[4979]: I1204 13:26:10.512104 4979 scope.go:117] "RemoveContainer" containerID="fd95c52414337c1a25dac76c57b06176c67d2a9903868ce8a22ddcfcec7f9a4b" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.513844 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.517457 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.519413 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.520384 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.549266 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.589666 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.589817 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-scripts\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.589863 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.589897 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-config-data\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.589965 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdzth\" (UniqueName: \"kubernetes.io/projected/ae3087b0-598c-46b6-a301-284d67b01ff6-kube-api-access-bdzth\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.590246 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-log-httpd\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.590371 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-run-httpd\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.692678 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-log-httpd\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.692740 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-run-httpd\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.692798 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.692882 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-scripts\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.692930 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.693042 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-config-data\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.693179 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdzth\" (UniqueName: \"kubernetes.io/projected/ae3087b0-598c-46b6-a301-284d67b01ff6-kube-api-access-bdzth\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.693242 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-log-httpd\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.693434 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-run-httpd\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.698851 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.699120 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.703013 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-scripts\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.707389 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-config-data\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.710786 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdzth\" (UniqueName: \"kubernetes.io/projected/ae3087b0-598c-46b6-a301-284d67b01ff6-kube-api-access-bdzth\") pod \"ceilometer-0\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " pod="openstack/ceilometer-0" Dec 04 13:26:11 crc kubenswrapper[4979]: I1204 13:26:11.838496 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.076660 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-bf9ff"] Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.089196 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-cj9ww"] Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.101890 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-bf9ff"] Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.113469 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9b7c-account-create-update-rb9m7"] Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.124433 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9b7c-account-create-update-rb9m7"] Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.134686 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-cj9ww"] Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.212013 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b71a0166-3ad9-47e6-9eec-7958a43af25c" path="/var/lib/kubelet/pods/b71a0166-3ad9-47e6-9eec-7958a43af25c/volumes" Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.214768 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc0a0593-111d-4346-a6e2-8831c97e3fd1" path="/var/lib/kubelet/pods/bc0a0593-111d-4346-a6e2-8831c97e3fd1/volumes" Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.216058 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca30de84-d450-49e5-909e-fc83c3dc09ca" path="/var/lib/kubelet/pods/ca30de84-d450-49e5-909e-fc83c3dc09ca/volumes" Dec 04 13:26:12 crc kubenswrapper[4979]: W1204 13:26:12.353647 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae3087b0_598c_46b6_a301_284d67b01ff6.slice/crio-be5982687422a59ae705a40221d5be82e875ae57e2114accf73dee9f884df448 WatchSource:0}: Error finding container be5982687422a59ae705a40221d5be82e875ae57e2114accf73dee9f884df448: Status 404 returned error can't find the container with id be5982687422a59ae705a40221d5be82e875ae57e2114accf73dee9f884df448 Dec 04 13:26:12 crc kubenswrapper[4979]: I1204 13:26:12.353918 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.033112 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-292c-account-create-update-8l69r"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.044722 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-292c-account-create-update-8l69r"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.056362 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-6407-account-create-update-zl277"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.065421 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-thbbq"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.076444 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-6407-account-create-update-zl277"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.091477 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-thbbq"] Dec 04 13:26:13 crc kubenswrapper[4979]: I1204 13:26:13.117002 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerStarted","Data":"be5982687422a59ae705a40221d5be82e875ae57e2114accf73dee9f884df448"} Dec 04 13:26:14 crc kubenswrapper[4979]: I1204 13:26:14.222408 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d3156be-8704-4b27-9cd9-1dd2b680d658" path="/var/lib/kubelet/pods/0d3156be-8704-4b27-9cd9-1dd2b680d658/volumes" Dec 04 13:26:14 crc kubenswrapper[4979]: I1204 13:26:14.226106 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5bc4316-e6bc-4ab5-8d69-d45305523691" path="/var/lib/kubelet/pods/e5bc4316-e6bc-4ab5-8d69-d45305523691/volumes" Dec 04 13:26:14 crc kubenswrapper[4979]: I1204 13:26:14.228794 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f691ee1e-f296-4817-bc2d-35d0b7b10f7d" path="/var/lib/kubelet/pods/f691ee1e-f296-4817-bc2d-35d0b7b10f7d/volumes" Dec 04 13:26:15 crc kubenswrapper[4979]: I1204 13:26:15.146204 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerStarted","Data":"575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b"} Dec 04 13:26:16 crc kubenswrapper[4979]: I1204 13:26:16.162032 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerStarted","Data":"98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a"} Dec 04 13:26:17 crc kubenswrapper[4979]: I1204 13:26:17.173841 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerStarted","Data":"cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82"} Dec 04 13:26:18 crc kubenswrapper[4979]: I1204 13:26:18.185733 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerStarted","Data":"ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73"} Dec 04 13:26:18 crc kubenswrapper[4979]: I1204 13:26:18.186297 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 13:26:18 crc kubenswrapper[4979]: I1204 13:26:18.222717 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.940285933 podStartE2EDuration="7.222696088s" podCreationTimestamp="2025-12-04 13:26:11 +0000 UTC" firstStartedPulling="2025-12-04 13:26:12.35614348 +0000 UTC m=+6196.630439284" lastFinishedPulling="2025-12-04 13:26:17.638553635 +0000 UTC m=+6201.912849439" observedRunningTime="2025-12-04 13:26:18.208612631 +0000 UTC m=+6202.482908435" watchObservedRunningTime="2025-12-04 13:26:18.222696088 +0000 UTC m=+6202.496991892" Dec 04 13:26:19 crc kubenswrapper[4979]: I1204 13:26:19.181513 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 04 13:26:19 crc kubenswrapper[4979]: I1204 13:26:19.185247 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 04 13:26:19 crc kubenswrapper[4979]: I1204 13:26:19.201286 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.051368 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rjk2s"] Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.067855 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-rjk2s"] Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.246700 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-rddkh"] Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.248419 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.257223 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-rddkh"] Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.349762 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-2e13-account-create-update-z9rxp"] Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.351762 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.353687 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.369933 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2e13-account-create-update-z9rxp"] Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.378868 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glxjl\" (UniqueName: \"kubernetes.io/projected/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-kube-api-access-glxjl\") pod \"aodh-db-create-rddkh\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.378957 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-operator-scripts\") pod \"aodh-db-create-rddkh\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.481326 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glxjl\" (UniqueName: \"kubernetes.io/projected/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-kube-api-access-glxjl\") pod \"aodh-db-create-rddkh\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.481387 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-operator-scripts\") pod \"aodh-db-create-rddkh\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.481542 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c77m8\" (UniqueName: \"kubernetes.io/projected/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-kube-api-access-c77m8\") pod \"aodh-2e13-account-create-update-z9rxp\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.481576 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-operator-scripts\") pod \"aodh-2e13-account-create-update-z9rxp\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.483489 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-operator-scripts\") pod \"aodh-db-create-rddkh\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.501953 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glxjl\" (UniqueName: \"kubernetes.io/projected/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-kube-api-access-glxjl\") pod \"aodh-db-create-rddkh\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.571567 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.583344 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c77m8\" (UniqueName: \"kubernetes.io/projected/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-kube-api-access-c77m8\") pod \"aodh-2e13-account-create-update-z9rxp\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.583459 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-operator-scripts\") pod \"aodh-2e13-account-create-update-z9rxp\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.584400 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-operator-scripts\") pod \"aodh-2e13-account-create-update-z9rxp\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.605586 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c77m8\" (UniqueName: \"kubernetes.io/projected/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-kube-api-access-c77m8\") pod \"aodh-2e13-account-create-update-z9rxp\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:23 crc kubenswrapper[4979]: I1204 13:26:23.671162 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.056288 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-rddkh"] Dec 04 13:26:24 crc kubenswrapper[4979]: W1204 13:26:24.068210 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99a4d56b_c6b2_4574_b5b6_e8cbe7938c4a.slice/crio-57d5f559b7d3e17c685e539aec94b6f0adf7c1aa15be6b1424af30f6b5ef3e92 WatchSource:0}: Error finding container 57d5f559b7d3e17c685e539aec94b6f0adf7c1aa15be6b1424af30f6b5ef3e92: Status 404 returned error can't find the container with id 57d5f559b7d3e17c685e539aec94b6f0adf7c1aa15be6b1424af30f6b5ef3e92 Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.211920 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f" path="/var/lib/kubelet/pods/c98e14f3-b82e-41e7-8850-8ad4bfcb4b1f/volumes" Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.253177 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-2e13-account-create-update-z9rxp"] Dec 04 13:26:24 crc kubenswrapper[4979]: W1204 13:26:24.262655 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod548a7b59_3465_4157_b1f7_ac66e3e7aa6a.slice/crio-1c8fb8ae72a25972616c8629b9e75b7f816d21ca4fe3377e2bcfa4a896ca59b3 WatchSource:0}: Error finding container 1c8fb8ae72a25972616c8629b9e75b7f816d21ca4fe3377e2bcfa4a896ca59b3: Status 404 returned error can't find the container with id 1c8fb8ae72a25972616c8629b9e75b7f816d21ca4fe3377e2bcfa4a896ca59b3 Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.314617 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2e13-account-create-update-z9rxp" event={"ID":"548a7b59-3465-4157-b1f7-ac66e3e7aa6a","Type":"ContainerStarted","Data":"1c8fb8ae72a25972616c8629b9e75b7f816d21ca4fe3377e2bcfa4a896ca59b3"} Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.315937 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rddkh" event={"ID":"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a","Type":"ContainerStarted","Data":"99d7f20537bca087ccb36855fb78791136e48ba785d0af1ea8d1a921f402a28d"} Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.315977 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rddkh" event={"ID":"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a","Type":"ContainerStarted","Data":"57d5f559b7d3e17c685e539aec94b6f0adf7c1aa15be6b1424af30f6b5ef3e92"} Dec 04 13:26:24 crc kubenswrapper[4979]: I1204 13:26:24.341500 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-rddkh" podStartSLOduration=1.34147732 podStartE2EDuration="1.34147732s" podCreationTimestamp="2025-12-04 13:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:26:24.327819363 +0000 UTC m=+6208.602115167" watchObservedRunningTime="2025-12-04 13:26:24.34147732 +0000 UTC m=+6208.615773124" Dec 04 13:26:25 crc kubenswrapper[4979]: I1204 13:26:25.326698 4979 generic.go:334] "Generic (PLEG): container finished" podID="548a7b59-3465-4157-b1f7-ac66e3e7aa6a" containerID="ced6b82964cee69bcb2fe7620c8f3061e5b13172a108dba3925934d8a6a23bd1" exitCode=0 Dec 04 13:26:25 crc kubenswrapper[4979]: I1204 13:26:25.326917 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2e13-account-create-update-z9rxp" event={"ID":"548a7b59-3465-4157-b1f7-ac66e3e7aa6a","Type":"ContainerDied","Data":"ced6b82964cee69bcb2fe7620c8f3061e5b13172a108dba3925934d8a6a23bd1"} Dec 04 13:26:25 crc kubenswrapper[4979]: I1204 13:26:25.328623 4979 generic.go:334] "Generic (PLEG): container finished" podID="99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" containerID="99d7f20537bca087ccb36855fb78791136e48ba785d0af1ea8d1a921f402a28d" exitCode=0 Dec 04 13:26:25 crc kubenswrapper[4979]: I1204 13:26:25.328657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rddkh" event={"ID":"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a","Type":"ContainerDied","Data":"99d7f20537bca087ccb36855fb78791136e48ba785d0af1ea8d1a921f402a28d"} Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.811205 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.825292 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.852506 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c77m8\" (UniqueName: \"kubernetes.io/projected/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-kube-api-access-c77m8\") pod \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.852570 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glxjl\" (UniqueName: \"kubernetes.io/projected/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-kube-api-access-glxjl\") pod \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.852712 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-operator-scripts\") pod \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\" (UID: \"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a\") " Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.852925 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-operator-scripts\") pod \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\" (UID: \"548a7b59-3465-4157-b1f7-ac66e3e7aa6a\") " Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.853817 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "548a7b59-3465-4157-b1f7-ac66e3e7aa6a" (UID: "548a7b59-3465-4157-b1f7-ac66e3e7aa6a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.854685 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" (UID: "99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.859089 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-kube-api-access-glxjl" (OuterVolumeSpecName: "kube-api-access-glxjl") pod "99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" (UID: "99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a"). InnerVolumeSpecName "kube-api-access-glxjl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.865515 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-kube-api-access-c77m8" (OuterVolumeSpecName: "kube-api-access-c77m8") pod "548a7b59-3465-4157-b1f7-ac66e3e7aa6a" (UID: "548a7b59-3465-4157-b1f7-ac66e3e7aa6a"). InnerVolumeSpecName "kube-api-access-c77m8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.956127 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c77m8\" (UniqueName: \"kubernetes.io/projected/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-kube-api-access-c77m8\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.956170 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glxjl\" (UniqueName: \"kubernetes.io/projected/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-kube-api-access-glxjl\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.956184 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:26 crc kubenswrapper[4979]: I1204 13:26:26.956196 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/548a7b59-3465-4157-b1f7-ac66e3e7aa6a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:27 crc kubenswrapper[4979]: I1204 13:26:27.362318 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-2e13-account-create-update-z9rxp" event={"ID":"548a7b59-3465-4157-b1f7-ac66e3e7aa6a","Type":"ContainerDied","Data":"1c8fb8ae72a25972616c8629b9e75b7f816d21ca4fe3377e2bcfa4a896ca59b3"} Dec 04 13:26:27 crc kubenswrapper[4979]: I1204 13:26:27.362787 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c8fb8ae72a25972616c8629b9e75b7f816d21ca4fe3377e2bcfa4a896ca59b3" Dec 04 13:26:27 crc kubenswrapper[4979]: I1204 13:26:27.362606 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-2e13-account-create-update-z9rxp" Dec 04 13:26:27 crc kubenswrapper[4979]: I1204 13:26:27.369394 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rddkh" event={"ID":"99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a","Type":"ContainerDied","Data":"57d5f559b7d3e17c685e539aec94b6f0adf7c1aa15be6b1424af30f6b5ef3e92"} Dec 04 13:26:27 crc kubenswrapper[4979]: I1204 13:26:27.369435 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57d5f559b7d3e17c685e539aec94b6f0adf7c1aa15be6b1424af30f6b5ef3e92" Dec 04 13:26:27 crc kubenswrapper[4979]: I1204 13:26:27.369509 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rddkh" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.734652 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-cvwdx"] Dec 04 13:26:28 crc kubenswrapper[4979]: E1204 13:26:28.735498 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" containerName="mariadb-database-create" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.735516 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" containerName="mariadb-database-create" Dec 04 13:26:28 crc kubenswrapper[4979]: E1204 13:26:28.735545 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="548a7b59-3465-4157-b1f7-ac66e3e7aa6a" containerName="mariadb-account-create-update" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.735553 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="548a7b59-3465-4157-b1f7-ac66e3e7aa6a" containerName="mariadb-account-create-update" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.735810 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="548a7b59-3465-4157-b1f7-ac66e3e7aa6a" containerName="mariadb-account-create-update" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.735831 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" containerName="mariadb-database-create" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.737166 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.740100 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.740436 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.742835 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.743835 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-zrk7s" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.752977 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-cvwdx"] Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.793704 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-config-data\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.793999 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqfj\" (UniqueName: \"kubernetes.io/projected/44910178-dbfd-4f8d-8e28-4357a1927185-kube-api-access-rkqfj\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.794169 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-combined-ca-bundle\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.794361 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-scripts\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.896860 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqfj\" (UniqueName: \"kubernetes.io/projected/44910178-dbfd-4f8d-8e28-4357a1927185-kube-api-access-rkqfj\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.896987 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-combined-ca-bundle\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.897086 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-scripts\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.897154 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-config-data\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.903685 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-scripts\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.908259 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-combined-ca-bundle\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.916271 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqfj\" (UniqueName: \"kubernetes.io/projected/44910178-dbfd-4f8d-8e28-4357a1927185-kube-api-access-rkqfj\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:28 crc kubenswrapper[4979]: I1204 13:26:28.918442 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-config-data\") pod \"aodh-db-sync-cvwdx\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:29 crc kubenswrapper[4979]: I1204 13:26:29.055626 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:29 crc kubenswrapper[4979]: I1204 13:26:29.553536 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-cvwdx"] Dec 04 13:26:30 crc kubenswrapper[4979]: I1204 13:26:30.403423 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-cvwdx" event={"ID":"44910178-dbfd-4f8d-8e28-4357a1927185","Type":"ContainerStarted","Data":"bf1c76792e9d995e341d1abe18988496338b5d18602f1239cfa8ab7372013a1d"} Dec 04 13:26:35 crc kubenswrapper[4979]: I1204 13:26:35.478998 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-cvwdx" event={"ID":"44910178-dbfd-4f8d-8e28-4357a1927185","Type":"ContainerStarted","Data":"1dbaac49665d193b86af97e3f435c947e6c70b729ece2244804081bea071060c"} Dec 04 13:26:35 crc kubenswrapper[4979]: I1204 13:26:35.514250 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-cvwdx" podStartSLOduration=2.724410883 podStartE2EDuration="7.51422696s" podCreationTimestamp="2025-12-04 13:26:28 +0000 UTC" firstStartedPulling="2025-12-04 13:26:29.564281945 +0000 UTC m=+6213.838577749" lastFinishedPulling="2025-12-04 13:26:34.354098022 +0000 UTC m=+6218.628393826" observedRunningTime="2025-12-04 13:26:35.4955575 +0000 UTC m=+6219.769853304" watchObservedRunningTime="2025-12-04 13:26:35.51422696 +0000 UTC m=+6219.788522764" Dec 04 13:26:37 crc kubenswrapper[4979]: I1204 13:26:37.499061 4979 generic.go:334] "Generic (PLEG): container finished" podID="44910178-dbfd-4f8d-8e28-4357a1927185" containerID="1dbaac49665d193b86af97e3f435c947e6c70b729ece2244804081bea071060c" exitCode=0 Dec 04 13:26:37 crc kubenswrapper[4979]: I1204 13:26:37.499153 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-cvwdx" event={"ID":"44910178-dbfd-4f8d-8e28-4357a1927185","Type":"ContainerDied","Data":"1dbaac49665d193b86af97e3f435c947e6c70b729ece2244804081bea071060c"} Dec 04 13:26:38 crc kubenswrapper[4979]: I1204 13:26:38.905712 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.020167 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rkqfj\" (UniqueName: \"kubernetes.io/projected/44910178-dbfd-4f8d-8e28-4357a1927185-kube-api-access-rkqfj\") pod \"44910178-dbfd-4f8d-8e28-4357a1927185\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.020464 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-scripts\") pod \"44910178-dbfd-4f8d-8e28-4357a1927185\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.020549 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-combined-ca-bundle\") pod \"44910178-dbfd-4f8d-8e28-4357a1927185\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.020573 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-config-data\") pod \"44910178-dbfd-4f8d-8e28-4357a1927185\" (UID: \"44910178-dbfd-4f8d-8e28-4357a1927185\") " Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.030796 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-scripts" (OuterVolumeSpecName: "scripts") pod "44910178-dbfd-4f8d-8e28-4357a1927185" (UID: "44910178-dbfd-4f8d-8e28-4357a1927185"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.030873 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44910178-dbfd-4f8d-8e28-4357a1927185-kube-api-access-rkqfj" (OuterVolumeSpecName: "kube-api-access-rkqfj") pod "44910178-dbfd-4f8d-8e28-4357a1927185" (UID: "44910178-dbfd-4f8d-8e28-4357a1927185"). InnerVolumeSpecName "kube-api-access-rkqfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.052875 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-config-data" (OuterVolumeSpecName: "config-data") pod "44910178-dbfd-4f8d-8e28-4357a1927185" (UID: "44910178-dbfd-4f8d-8e28-4357a1927185"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.055710 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "44910178-dbfd-4f8d-8e28-4357a1927185" (UID: "44910178-dbfd-4f8d-8e28-4357a1927185"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.122731 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.122756 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.122767 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/44910178-dbfd-4f8d-8e28-4357a1927185-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.122776 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rkqfj\" (UniqueName: \"kubernetes.io/projected/44910178-dbfd-4f8d-8e28-4357a1927185-kube-api-access-rkqfj\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.521078 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-cvwdx" event={"ID":"44910178-dbfd-4f8d-8e28-4357a1927185","Type":"ContainerDied","Data":"bf1c76792e9d995e341d1abe18988496338b5d18602f1239cfa8ab7372013a1d"} Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.521124 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf1c76792e9d995e341d1abe18988496338b5d18602f1239cfa8ab7372013a1d" Dec 04 13:26:39 crc kubenswrapper[4979]: I1204 13:26:39.521129 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-cvwdx" Dec 04 13:26:41 crc kubenswrapper[4979]: I1204 13:26:41.029427 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5c8l"] Dec 04 13:26:41 crc kubenswrapper[4979]: I1204 13:26:41.049691 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-s5c8l"] Dec 04 13:26:41 crc kubenswrapper[4979]: I1204 13:26:41.846108 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 13:26:42 crc kubenswrapper[4979]: I1204 13:26:42.214865 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6741056-1468-4cc3-b55b-af86b1702693" path="/var/lib/kubelet/pods/e6741056-1468-4cc3-b55b-af86b1702693/volumes" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.039613 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-ljxl8"] Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.048975 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-ljxl8"] Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.953758 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 04 13:26:43 crc kubenswrapper[4979]: E1204 13:26:43.954455 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44910178-dbfd-4f8d-8e28-4357a1927185" containerName="aodh-db-sync" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.954468 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="44910178-dbfd-4f8d-8e28-4357a1927185" containerName="aodh-db-sync" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.954675 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="44910178-dbfd-4f8d-8e28-4357a1927185" containerName="aodh-db-sync" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.959788 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.962228 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-zrk7s" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.965542 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.965963 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 04 13:26:43 crc kubenswrapper[4979]: I1204 13:26:43.973975 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.038411 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4glm\" (UniqueName: \"kubernetes.io/projected/36d65388-9f95-4763-afb2-57c66bbc47e2-kube-api-access-h4glm\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.038692 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-scripts\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.038905 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.039137 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-config-data\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.143558 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.143689 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-config-data\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.143740 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4glm\" (UniqueName: \"kubernetes.io/projected/36d65388-9f95-4763-afb2-57c66bbc47e2-kube-api-access-h4glm\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.143824 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-scripts\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.153703 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-config-data\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.154291 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-combined-ca-bundle\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.163025 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/36d65388-9f95-4763-afb2-57c66bbc47e2-scripts\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.177456 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4glm\" (UniqueName: \"kubernetes.io/projected/36d65388-9f95-4763-afb2-57c66bbc47e2-kube-api-access-h4glm\") pod \"aodh-0\" (UID: \"36d65388-9f95-4763-afb2-57c66bbc47e2\") " pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.218275 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d9fb02d-ba21-43b8-ba02-5d15d566d01e" path="/var/lib/kubelet/pods/4d9fb02d-ba21-43b8-ba02-5d15d566d01e/volumes" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.296221 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 04 13:26:44 crc kubenswrapper[4979]: I1204 13:26:44.819267 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.100812 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.101067 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-central-agent" containerID="cri-o://575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b" gracePeriod=30 Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.101122 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="proxy-httpd" containerID="cri-o://ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73" gracePeriod=30 Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.101189 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="sg-core" containerID="cri-o://cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82" gracePeriod=30 Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.101238 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-notification-agent" containerID="cri-o://98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a" gracePeriod=30 Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.594064 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"36d65388-9f95-4763-afb2-57c66bbc47e2","Type":"ContainerStarted","Data":"3234855c7cb2fc546d37ed5982385865d999056493a6358893b1eb12e307d65d"} Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.611835 4979 generic.go:334] "Generic (PLEG): container finished" podID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerID="ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73" exitCode=0 Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.611879 4979 generic.go:334] "Generic (PLEG): container finished" podID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerID="cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82" exitCode=2 Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.611929 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerDied","Data":"ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73"} Dec 04 13:26:45 crc kubenswrapper[4979]: I1204 13:26:45.612005 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerDied","Data":"cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82"} Dec 04 13:26:46 crc kubenswrapper[4979]: I1204 13:26:46.625964 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"36d65388-9f95-4763-afb2-57c66bbc47e2","Type":"ContainerStarted","Data":"0e4b6164dea7b8b1933a7a391d3ceb846b744f7ac0ecabbf104dcc2ebe590db5"} Dec 04 13:26:46 crc kubenswrapper[4979]: I1204 13:26:46.631189 4979 generic.go:334] "Generic (PLEG): container finished" podID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerID="575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b" exitCode=0 Dec 04 13:26:46 crc kubenswrapper[4979]: I1204 13:26:46.631249 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerDied","Data":"575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b"} Dec 04 13:26:47 crc kubenswrapper[4979]: I1204 13:26:47.649758 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"36d65388-9f95-4763-afb2-57c66bbc47e2","Type":"ContainerStarted","Data":"f3344cd534418959b127f1f9f5bea16047152f32b9d94fbaa4c073b7b6bd027e"} Dec 04 13:26:48 crc kubenswrapper[4979]: I1204 13:26:48.664621 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"36d65388-9f95-4763-afb2-57c66bbc47e2","Type":"ContainerStarted","Data":"26df18ac9408778789d9bf050ddaa9991585e2738eca82eea5badc7ac9e0b6c2"} Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.292178 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.371405 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdzth\" (UniqueName: \"kubernetes.io/projected/ae3087b0-598c-46b6-a301-284d67b01ff6-kube-api-access-bdzth\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.371984 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-sg-core-conf-yaml\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.372026 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-run-httpd\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.372065 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-log-httpd\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.372180 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-scripts\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.372213 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-config-data\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.372648 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-combined-ca-bundle\") pod \"ae3087b0-598c-46b6-a301-284d67b01ff6\" (UID: \"ae3087b0-598c-46b6-a301-284d67b01ff6\") " Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.374015 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.380536 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3087b0-598c-46b6-a301-284d67b01ff6-kube-api-access-bdzth" (OuterVolumeSpecName: "kube-api-access-bdzth") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "kube-api-access-bdzth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.382274 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.391549 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-scripts" (OuterVolumeSpecName: "scripts") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.414594 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.475122 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdzth\" (UniqueName: \"kubernetes.io/projected/ae3087b0-598c-46b6-a301-284d67b01ff6-kube-api-access-bdzth\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.475164 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.475181 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.475191 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae3087b0-598c-46b6-a301-284d67b01ff6-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.475201 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.483413 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.490681 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-config-data" (OuterVolumeSpecName: "config-data") pod "ae3087b0-598c-46b6-a301-284d67b01ff6" (UID: "ae3087b0-598c-46b6-a301-284d67b01ff6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.577102 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.577183 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3087b0-598c-46b6-a301-284d67b01ff6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.679131 4979 generic.go:334] "Generic (PLEG): container finished" podID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerID="98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a" exitCode=0 Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.679174 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerDied","Data":"98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a"} Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.679202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae3087b0-598c-46b6-a301-284d67b01ff6","Type":"ContainerDied","Data":"be5982687422a59ae705a40221d5be82e875ae57e2114accf73dee9f884df448"} Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.679220 4979 scope.go:117] "RemoveContainer" containerID="ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.679261 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.705980 4979 scope.go:117] "RemoveContainer" containerID="cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.729934 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.747962 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.748826 4979 scope.go:117] "RemoveContainer" containerID="98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.763142 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.763767 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-central-agent" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.763795 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-central-agent" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.763813 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-notification-agent" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.763826 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-notification-agent" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.763857 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="sg-core" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.763872 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="sg-core" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.763929 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="proxy-httpd" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.763941 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="proxy-httpd" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.764274 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="proxy-httpd" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.764490 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-central-agent" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.764545 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="sg-core" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.764565 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" containerName="ceilometer-notification-agent" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.769835 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.774725 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.775053 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.778688 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.803588 4979 scope.go:117] "RemoveContainer" containerID="575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.837464 4979 scope.go:117] "RemoveContainer" containerID="ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.838218 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73\": container with ID starting with ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73 not found: ID does not exist" containerID="ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.838260 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73"} err="failed to get container status \"ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73\": rpc error: code = NotFound desc = could not find container \"ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73\": container with ID starting with ddd8b9ed1e0b1d2c6a28f485f34a661743d69dec74b75971f4f45d50462ebf73 not found: ID does not exist" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.838290 4979 scope.go:117] "RemoveContainer" containerID="cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.839379 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82\": container with ID starting with cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82 not found: ID does not exist" containerID="cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.839403 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82"} err="failed to get container status \"cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82\": rpc error: code = NotFound desc = could not find container \"cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82\": container with ID starting with cb0924f2692e34fcf13b254647ae331614f1769f5fdc75c37de062896e06ea82 not found: ID does not exist" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.839422 4979 scope.go:117] "RemoveContainer" containerID="98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.840134 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a\": container with ID starting with 98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a not found: ID does not exist" containerID="98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.840229 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a"} err="failed to get container status \"98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a\": rpc error: code = NotFound desc = could not find container \"98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a\": container with ID starting with 98173ef5acd84e2c384f98d4682c59ee3c7395df55e96ff506ab96fb0991b18a not found: ID does not exist" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.840287 4979 scope.go:117] "RemoveContainer" containerID="575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b" Dec 04 13:26:49 crc kubenswrapper[4979]: E1204 13:26:49.840830 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b\": container with ID starting with 575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b not found: ID does not exist" containerID="575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.840932 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b"} err="failed to get container status \"575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b\": rpc error: code = NotFound desc = could not find container \"575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b\": container with ID starting with 575a7b52c05d8028fae67590a9cf2617cc56f718c8908ed52eec9e032a6e735b not found: ID does not exist" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.882808 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-run-httpd\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.882897 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-config-data\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.883269 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-log-httpd\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.883467 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.883742 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.883924 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-scripts\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.883954 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r8k7\" (UniqueName: \"kubernetes.io/projected/15ef03c7-1056-4423-bb89-1ea324bbd3be-kube-api-access-9r8k7\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986094 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986181 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986266 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-scripts\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986330 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r8k7\" (UniqueName: \"kubernetes.io/projected/15ef03c7-1056-4423-bb89-1ea324bbd3be-kube-api-access-9r8k7\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986389 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-run-httpd\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986455 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-config-data\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.986511 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-log-httpd\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.987250 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-run-httpd\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.987342 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-log-httpd\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.991989 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-config-data\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.992705 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.993128 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-scripts\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:49 crc kubenswrapper[4979]: I1204 13:26:49.997014 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:50 crc kubenswrapper[4979]: I1204 13:26:50.008733 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r8k7\" (UniqueName: \"kubernetes.io/projected/15ef03c7-1056-4423-bb89-1ea324bbd3be-kube-api-access-9r8k7\") pod \"ceilometer-0\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " pod="openstack/ceilometer-0" Dec 04 13:26:50 crc kubenswrapper[4979]: I1204 13:26:50.102754 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:26:50 crc kubenswrapper[4979]: I1204 13:26:50.214569 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3087b0-598c-46b6-a301-284d67b01ff6" path="/var/lib/kubelet/pods/ae3087b0-598c-46b6-a301-284d67b01ff6/volumes" Dec 04 13:26:50 crc kubenswrapper[4979]: I1204 13:26:50.694266 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"36d65388-9f95-4763-afb2-57c66bbc47e2","Type":"ContainerStarted","Data":"3eb70a4b2ab6b036f305b3a5c33bb8c18f6b5cec4bf1326e16bd89d94a8be7f2"} Dec 04 13:26:50 crc kubenswrapper[4979]: I1204 13:26:50.711201 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:26:50 crc kubenswrapper[4979]: W1204 13:26:50.716871 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15ef03c7_1056_4423_bb89_1ea324bbd3be.slice/crio-7226259a55a84cc4c582e2babf823fc407d1896558e32a53ce297d19861a3139 WatchSource:0}: Error finding container 7226259a55a84cc4c582e2babf823fc407d1896558e32a53ce297d19861a3139: Status 404 returned error can't find the container with id 7226259a55a84cc4c582e2babf823fc407d1896558e32a53ce297d19861a3139 Dec 04 13:26:50 crc kubenswrapper[4979]: I1204 13:26:50.725702 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.32610884 podStartE2EDuration="7.725680736s" podCreationTimestamp="2025-12-04 13:26:43 +0000 UTC" firstStartedPulling="2025-12-04 13:26:44.810795862 +0000 UTC m=+6229.085091666" lastFinishedPulling="2025-12-04 13:26:50.210367748 +0000 UTC m=+6234.484663562" observedRunningTime="2025-12-04 13:26:50.721257497 +0000 UTC m=+6234.995553351" watchObservedRunningTime="2025-12-04 13:26:50.725680736 +0000 UTC m=+6234.999976540" Dec 04 13:26:51 crc kubenswrapper[4979]: I1204 13:26:51.706951 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerStarted","Data":"7226259a55a84cc4c582e2babf823fc407d1896558e32a53ce297d19861a3139"} Dec 04 13:26:52 crc kubenswrapper[4979]: I1204 13:26:52.721871 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerStarted","Data":"2d2e7536d779e8b98085c87e5c3009706b53d62e29334f38404154b233f63f23"} Dec 04 13:26:53 crc kubenswrapper[4979]: I1204 13:26:53.735013 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerStarted","Data":"ad156b62041007ce7aef25003b37b0fc7768765fbae7d6e6d427c0bb29695aa6"} Dec 04 13:26:53 crc kubenswrapper[4979]: I1204 13:26:53.735379 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerStarted","Data":"ae8391deaa4546d7f0ccc9c955c02efe5e501c91c6d7359d75693c84a5b6c8b9"} Dec 04 13:26:55 crc kubenswrapper[4979]: I1204 13:26:55.767106 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerStarted","Data":"c60543d267b30eb046a7935f6576af57b1774f574a5c7cd49274e51f8e8144e1"} Dec 04 13:26:55 crc kubenswrapper[4979]: I1204 13:26:55.767991 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 13:26:55 crc kubenswrapper[4979]: I1204 13:26:55.806792 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.951891256 podStartE2EDuration="6.806757482s" podCreationTimestamp="2025-12-04 13:26:49 +0000 UTC" firstStartedPulling="2025-12-04 13:26:50.718823162 +0000 UTC m=+6234.993118986" lastFinishedPulling="2025-12-04 13:26:54.573689418 +0000 UTC m=+6238.847985212" observedRunningTime="2025-12-04 13:26:55.803535806 +0000 UTC m=+6240.077831640" watchObservedRunningTime="2025-12-04 13:26:55.806757482 +0000 UTC m=+6240.081053336" Dec 04 13:26:56 crc kubenswrapper[4979]: I1204 13:26:56.969423 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-cwpc4"] Dec 04 13:26:56 crc kubenswrapper[4979]: I1204 13:26:56.972979 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.040841 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-cwpc4"] Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.046162 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140ae347-58c1-49a2-9ff1-9eb734203a83-operator-scripts\") pod \"manila-db-create-cwpc4\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.046609 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m24rj\" (UniqueName: \"kubernetes.io/projected/140ae347-58c1-49a2-9ff1-9eb734203a83-kube-api-access-m24rj\") pod \"manila-db-create-cwpc4\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.149051 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-ce1f-account-create-update-lkvfb"] Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.149706 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140ae347-58c1-49a2-9ff1-9eb734203a83-operator-scripts\") pod \"manila-db-create-cwpc4\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.149819 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m24rj\" (UniqueName: \"kubernetes.io/projected/140ae347-58c1-49a2-9ff1-9eb734203a83-kube-api-access-m24rj\") pod \"manila-db-create-cwpc4\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.150817 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.150878 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140ae347-58c1-49a2-9ff1-9eb734203a83-operator-scripts\") pod \"manila-db-create-cwpc4\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.154178 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.160421 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ce1f-account-create-update-lkvfb"] Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.192690 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m24rj\" (UniqueName: \"kubernetes.io/projected/140ae347-58c1-49a2-9ff1-9eb734203a83-kube-api-access-m24rj\") pod \"manila-db-create-cwpc4\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.253656 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxwxp\" (UniqueName: \"kubernetes.io/projected/429bb3c3-eb68-4baf-9ab1-380fcfa50351-kube-api-access-sxwxp\") pod \"manila-ce1f-account-create-update-lkvfb\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.254594 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/429bb3c3-eb68-4baf-9ab1-380fcfa50351-operator-scripts\") pod \"manila-ce1f-account-create-update-lkvfb\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.312245 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cwpc4" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.361196 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxwxp\" (UniqueName: \"kubernetes.io/projected/429bb3c3-eb68-4baf-9ab1-380fcfa50351-kube-api-access-sxwxp\") pod \"manila-ce1f-account-create-update-lkvfb\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.362367 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/429bb3c3-eb68-4baf-9ab1-380fcfa50351-operator-scripts\") pod \"manila-ce1f-account-create-update-lkvfb\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.363485 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/429bb3c3-eb68-4baf-9ab1-380fcfa50351-operator-scripts\") pod \"manila-ce1f-account-create-update-lkvfb\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.389998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxwxp\" (UniqueName: \"kubernetes.io/projected/429bb3c3-eb68-4baf-9ab1-380fcfa50351-kube-api-access-sxwxp\") pod \"manila-ce1f-account-create-update-lkvfb\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.479706 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:26:57 crc kubenswrapper[4979]: W1204 13:26:57.929526 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod140ae347_58c1_49a2_9ff1_9eb734203a83.slice/crio-3473cfad96aafb140043ba215bbbe2839188b94b07943ae3f2caffb8809619e2 WatchSource:0}: Error finding container 3473cfad96aafb140043ba215bbbe2839188b94b07943ae3f2caffb8809619e2: Status 404 returned error can't find the container with id 3473cfad96aafb140043ba215bbbe2839188b94b07943ae3f2caffb8809619e2 Dec 04 13:26:57 crc kubenswrapper[4979]: I1204 13:26:57.930927 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-cwpc4"] Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.084590 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-ce1f-account-create-update-lkvfb"] Dec 04 13:26:58 crc kubenswrapper[4979]: W1204 13:26:58.091406 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod429bb3c3_eb68_4baf_9ab1_380fcfa50351.slice/crio-b24561e744224ae5bc8640ec6f8edcd674e91a21b9300fdad2aec5936492988d WatchSource:0}: Error finding container b24561e744224ae5bc8640ec6f8edcd674e91a21b9300fdad2aec5936492988d: Status 404 returned error can't find the container with id b24561e744224ae5bc8640ec6f8edcd674e91a21b9300fdad2aec5936492988d Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.811641 4979 generic.go:334] "Generic (PLEG): container finished" podID="140ae347-58c1-49a2-9ff1-9eb734203a83" containerID="cb325e1503e264ba4e72615bfe4b6bf6c3027229f31d700e663bac7437836313" exitCode=0 Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.811679 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-cwpc4" event={"ID":"140ae347-58c1-49a2-9ff1-9eb734203a83","Type":"ContainerDied","Data":"cb325e1503e264ba4e72615bfe4b6bf6c3027229f31d700e663bac7437836313"} Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.811974 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-cwpc4" event={"ID":"140ae347-58c1-49a2-9ff1-9eb734203a83","Type":"ContainerStarted","Data":"3473cfad96aafb140043ba215bbbe2839188b94b07943ae3f2caffb8809619e2"} Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.813522 4979 generic.go:334] "Generic (PLEG): container finished" podID="429bb3c3-eb68-4baf-9ab1-380fcfa50351" containerID="0509d520f85761adb2c0cd1029774dc64f537e2bc2c34f0d2081c76c6591af52" exitCode=0 Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.813547 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ce1f-account-create-update-lkvfb" event={"ID":"429bb3c3-eb68-4baf-9ab1-380fcfa50351","Type":"ContainerDied","Data":"0509d520f85761adb2c0cd1029774dc64f537e2bc2c34f0d2081c76c6591af52"} Dec 04 13:26:58 crc kubenswrapper[4979]: I1204 13:26:58.813578 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ce1f-account-create-update-lkvfb" event={"ID":"429bb3c3-eb68-4baf-9ab1-380fcfa50351","Type":"ContainerStarted","Data":"b24561e744224ae5bc8640ec6f8edcd674e91a21b9300fdad2aec5936492988d"} Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.078177 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-gpkvf"] Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.090983 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-gpkvf"] Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.214154 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9f3c3d0-6cfd-4b30-b836-62671e12f561" path="/var/lib/kubelet/pods/f9f3c3d0-6cfd-4b30-b836-62671e12f561/volumes" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.291339 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cwpc4" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.299961 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.341864 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/429bb3c3-eb68-4baf-9ab1-380fcfa50351-operator-scripts\") pod \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.342054 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140ae347-58c1-49a2-9ff1-9eb734203a83-operator-scripts\") pod \"140ae347-58c1-49a2-9ff1-9eb734203a83\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.342235 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m24rj\" (UniqueName: \"kubernetes.io/projected/140ae347-58c1-49a2-9ff1-9eb734203a83-kube-api-access-m24rj\") pod \"140ae347-58c1-49a2-9ff1-9eb734203a83\" (UID: \"140ae347-58c1-49a2-9ff1-9eb734203a83\") " Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.342509 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxwxp\" (UniqueName: \"kubernetes.io/projected/429bb3c3-eb68-4baf-9ab1-380fcfa50351-kube-api-access-sxwxp\") pod \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\" (UID: \"429bb3c3-eb68-4baf-9ab1-380fcfa50351\") " Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.344484 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/429bb3c3-eb68-4baf-9ab1-380fcfa50351-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "429bb3c3-eb68-4baf-9ab1-380fcfa50351" (UID: "429bb3c3-eb68-4baf-9ab1-380fcfa50351"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.344489 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/140ae347-58c1-49a2-9ff1-9eb734203a83-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "140ae347-58c1-49a2-9ff1-9eb734203a83" (UID: "140ae347-58c1-49a2-9ff1-9eb734203a83"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.348234 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/429bb3c3-eb68-4baf-9ab1-380fcfa50351-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.348273 4979 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/140ae347-58c1-49a2-9ff1-9eb734203a83-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.351192 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/429bb3c3-eb68-4baf-9ab1-380fcfa50351-kube-api-access-sxwxp" (OuterVolumeSpecName: "kube-api-access-sxwxp") pod "429bb3c3-eb68-4baf-9ab1-380fcfa50351" (UID: "429bb3c3-eb68-4baf-9ab1-380fcfa50351"). InnerVolumeSpecName "kube-api-access-sxwxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.363429 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/140ae347-58c1-49a2-9ff1-9eb734203a83-kube-api-access-m24rj" (OuterVolumeSpecName: "kube-api-access-m24rj") pod "140ae347-58c1-49a2-9ff1-9eb734203a83" (UID: "140ae347-58c1-49a2-9ff1-9eb734203a83"). InnerVolumeSpecName "kube-api-access-m24rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.452172 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m24rj\" (UniqueName: \"kubernetes.io/projected/140ae347-58c1-49a2-9ff1-9eb734203a83-kube-api-access-m24rj\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.452234 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxwxp\" (UniqueName: \"kubernetes.io/projected/429bb3c3-eb68-4baf-9ab1-380fcfa50351-kube-api-access-sxwxp\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.835352 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-cwpc4" event={"ID":"140ae347-58c1-49a2-9ff1-9eb734203a83","Type":"ContainerDied","Data":"3473cfad96aafb140043ba215bbbe2839188b94b07943ae3f2caffb8809619e2"} Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.835413 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3473cfad96aafb140043ba215bbbe2839188b94b07943ae3f2caffb8809619e2" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.835502 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-cwpc4" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.851348 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-ce1f-account-create-update-lkvfb" event={"ID":"429bb3c3-eb68-4baf-9ab1-380fcfa50351","Type":"ContainerDied","Data":"b24561e744224ae5bc8640ec6f8edcd674e91a21b9300fdad2aec5936492988d"} Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.851402 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b24561e744224ae5bc8640ec6f8edcd674e91a21b9300fdad2aec5936492988d" Dec 04 13:27:00 crc kubenswrapper[4979]: I1204 13:27:00.851479 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-ce1f-account-create-update-lkvfb" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.498107 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-ztz4s"] Dec 04 13:27:02 crc kubenswrapper[4979]: E1204 13:27:02.499109 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="429bb3c3-eb68-4baf-9ab1-380fcfa50351" containerName="mariadb-account-create-update" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.499130 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="429bb3c3-eb68-4baf-9ab1-380fcfa50351" containerName="mariadb-account-create-update" Dec 04 13:27:02 crc kubenswrapper[4979]: E1204 13:27:02.499175 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="140ae347-58c1-49a2-9ff1-9eb734203a83" containerName="mariadb-database-create" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.499184 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="140ae347-58c1-49a2-9ff1-9eb734203a83" containerName="mariadb-database-create" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.499518 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="429bb3c3-eb68-4baf-9ab1-380fcfa50351" containerName="mariadb-account-create-update" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.499545 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="140ae347-58c1-49a2-9ff1-9eb734203a83" containerName="mariadb-database-create" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.500557 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.505397 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.508266 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-mwqlv" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.510453 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ztz4s"] Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.593879 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-config-data\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.594135 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-job-config-data\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.594279 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-combined-ca-bundle\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.594524 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrkvq\" (UniqueName: \"kubernetes.io/projected/0a57214b-fa12-4ac5-bc3a-773b80a03a19-kube-api-access-wrkvq\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.696893 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-job-config-data\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.696957 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-combined-ca-bundle\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.697026 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrkvq\" (UniqueName: \"kubernetes.io/projected/0a57214b-fa12-4ac5-bc3a-773b80a03a19-kube-api-access-wrkvq\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.697084 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-config-data\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.703277 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-job-config-data\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.703628 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-config-data\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.704098 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-combined-ca-bundle\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.714866 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrkvq\" (UniqueName: \"kubernetes.io/projected/0a57214b-fa12-4ac5-bc3a-773b80a03a19-kube-api-access-wrkvq\") pod \"manila-db-sync-ztz4s\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:02 crc kubenswrapper[4979]: I1204 13:27:02.834402 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:03 crc kubenswrapper[4979]: I1204 13:27:03.574368 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-ztz4s"] Dec 04 13:27:03 crc kubenswrapper[4979]: W1204 13:27:03.576000 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a57214b_fa12_4ac5_bc3a_773b80a03a19.slice/crio-094644d38d0d0e310357348f4819f37992085f8f22221310abe68aeb3476b712 WatchSource:0}: Error finding container 094644d38d0d0e310357348f4819f37992085f8f22221310abe68aeb3476b712: Status 404 returned error can't find the container with id 094644d38d0d0e310357348f4819f37992085f8f22221310abe68aeb3476b712 Dec 04 13:27:03 crc kubenswrapper[4979]: I1204 13:27:03.881810 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ztz4s" event={"ID":"0a57214b-fa12-4ac5-bc3a-773b80a03a19","Type":"ContainerStarted","Data":"094644d38d0d0e310357348f4819f37992085f8f22221310abe68aeb3476b712"} Dec 04 13:27:08 crc kubenswrapper[4979]: I1204 13:27:08.933858 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ztz4s" event={"ID":"0a57214b-fa12-4ac5-bc3a-773b80a03a19","Type":"ContainerStarted","Data":"7d7fe9d7db5308742dd3e78daab1f0a1d44d39e2c67b0e6c62d6a1d5ed8dc40d"} Dec 04 13:27:08 crc kubenswrapper[4979]: I1204 13:27:08.984765 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-ztz4s" podStartSLOduration=3.161680638 podStartE2EDuration="6.984744752s" podCreationTimestamp="2025-12-04 13:27:02 +0000 UTC" firstStartedPulling="2025-12-04 13:27:03.579574905 +0000 UTC m=+6247.853870709" lastFinishedPulling="2025-12-04 13:27:07.402639019 +0000 UTC m=+6251.676934823" observedRunningTime="2025-12-04 13:27:08.946984459 +0000 UTC m=+6253.221280273" watchObservedRunningTime="2025-12-04 13:27:08.984744752 +0000 UTC m=+6253.259040556" Dec 04 13:27:09 crc kubenswrapper[4979]: I1204 13:27:09.944752 4979 generic.go:334] "Generic (PLEG): container finished" podID="0a57214b-fa12-4ac5-bc3a-773b80a03a19" containerID="7d7fe9d7db5308742dd3e78daab1f0a1d44d39e2c67b0e6c62d6a1d5ed8dc40d" exitCode=0 Dec 04 13:27:09 crc kubenswrapper[4979]: I1204 13:27:09.944808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ztz4s" event={"ID":"0a57214b-fa12-4ac5-bc3a-773b80a03a19","Type":"ContainerDied","Data":"7d7fe9d7db5308742dd3e78daab1f0a1d44d39e2c67b0e6c62d6a1d5ed8dc40d"} Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.665452 4979 scope.go:117] "RemoveContainer" containerID="3a3976ec5a3960ebe0310455ad3614b0c8ee93d0c921c736d3145b874d10140d" Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.689576 4979 scope.go:117] "RemoveContainer" containerID="2f32354589cfbd41f44ba0945614c186d20eb3b87d26b5b699fa1e67ab9eb088" Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.747897 4979 scope.go:117] "RemoveContainer" containerID="db71e79090ed41fbcf86fe025d7f936cca867a9118e67fc3b1d09dc2e175fa87" Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.804743 4979 scope.go:117] "RemoveContainer" containerID="b44a1ad0450092e02a5bbaea44d48fa05dc01d8d7df47246441925799ab8bfb4" Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.874369 4979 scope.go:117] "RemoveContainer" containerID="60dabd3bd55f05b03b087d547575cdd9816874a5885fc00f6b02fd600defe8f3" Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.905063 4979 scope.go:117] "RemoveContainer" containerID="83c1378a964ea7d546fb7e5044effc3da6d6be1e52c9ece03f3b751507a29b41" Dec 04 13:27:10 crc kubenswrapper[4979]: I1204 13:27:10.975783 4979 scope.go:117] "RemoveContainer" containerID="4caf6195a47109467c3b03032b264b2c37dd8d6870341f757c302d8879ecfd19" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.025684 4979 scope.go:117] "RemoveContainer" containerID="e4bef2e6aea11e141ed1f01fed1205e41bd91b8727c00e92faf7dba24963fae9" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.069876 4979 scope.go:117] "RemoveContainer" containerID="8377b5b30088979b3a06fa5bb98286f97bf7c2e494558f84627ebb018bbf8c03" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.095950 4979 scope.go:117] "RemoveContainer" containerID="cab696639147898acb787288da495c7dd9c1889282a9371c77926c223c265eb2" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.115687 4979 scope.go:117] "RemoveContainer" containerID="3b3b9a4fc820fbd1b5fd775bf777c40ce6981214ad55407d855291c8db502be8" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.374066 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.511199 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-job-config-data\") pod \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.511279 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-combined-ca-bundle\") pod \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.511320 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-config-data\") pod \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.511528 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrkvq\" (UniqueName: \"kubernetes.io/projected/0a57214b-fa12-4ac5-bc3a-773b80a03a19-kube-api-access-wrkvq\") pod \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\" (UID: \"0a57214b-fa12-4ac5-bc3a-773b80a03a19\") " Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.517692 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "0a57214b-fa12-4ac5-bc3a-773b80a03a19" (UID: "0a57214b-fa12-4ac5-bc3a-773b80a03a19"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.518061 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a57214b-fa12-4ac5-bc3a-773b80a03a19-kube-api-access-wrkvq" (OuterVolumeSpecName: "kube-api-access-wrkvq") pod "0a57214b-fa12-4ac5-bc3a-773b80a03a19" (UID: "0a57214b-fa12-4ac5-bc3a-773b80a03a19"). InnerVolumeSpecName "kube-api-access-wrkvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.520061 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-config-data" (OuterVolumeSpecName: "config-data") pod "0a57214b-fa12-4ac5-bc3a-773b80a03a19" (UID: "0a57214b-fa12-4ac5-bc3a-773b80a03a19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.542496 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a57214b-fa12-4ac5-bc3a-773b80a03a19" (UID: "0a57214b-fa12-4ac5-bc3a-773b80a03a19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.614504 4979 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.615441 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.615547 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a57214b-fa12-4ac5-bc3a-773b80a03a19-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.615636 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrkvq\" (UniqueName: \"kubernetes.io/projected/0a57214b-fa12-4ac5-bc3a-773b80a03a19-kube-api-access-wrkvq\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.993726 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-ztz4s" event={"ID":"0a57214b-fa12-4ac5-bc3a-773b80a03a19","Type":"ContainerDied","Data":"094644d38d0d0e310357348f4819f37992085f8f22221310abe68aeb3476b712"} Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.994072 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="094644d38d0d0e310357348f4819f37992085f8f22221310abe68aeb3476b712" Dec 04 13:27:11 crc kubenswrapper[4979]: I1204 13:27:11.993847 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-ztz4s" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.234110 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 04 13:27:12 crc kubenswrapper[4979]: E1204 13:27:12.234719 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a57214b-fa12-4ac5-bc3a-773b80a03a19" containerName="manila-db-sync" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.234741 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a57214b-fa12-4ac5-bc3a-773b80a03a19" containerName="manila-db-sync" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.235039 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a57214b-fa12-4ac5-bc3a-773b80a03a19" containerName="manila-db-sync" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.236489 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.240699 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.240864 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.242350 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-mwqlv" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.246484 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.249044 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.249905 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.254824 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.272223 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.305887 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329662 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-config-data\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/defcd96a-640e-44a5-881e-b45711312d7e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329734 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329781 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329797 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-config-data\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329862 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-ceph\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329887 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-scripts\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329928 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/defcd96a-640e-44a5-881e-b45711312d7e-kube-api-access-fcs8c\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329956 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.329974 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.330003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.330019 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-scripts\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.330037 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4b9hq\" (UniqueName: \"kubernetes.io/projected/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-kube-api-access-4b9hq\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.330055 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.402718 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cfb6864c7-7wk4c"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.412875 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.422382 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cfb6864c7-7wk4c"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.431834 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-config\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.431903 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.433664 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-config-data\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.433742 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtxz4\" (UniqueName: \"kubernetes.io/projected/1e98e71b-d235-4b50-8145-a445b8b5db8c-kube-api-access-gtxz4\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.433858 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-dns-svc\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434119 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-ceph\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434160 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-scripts\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434197 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-nb\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434287 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/defcd96a-640e-44a5-881e-b45711312d7e-kube-api-access-fcs8c\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434382 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434420 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434507 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434544 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-scripts\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434595 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4b9hq\" (UniqueName: \"kubernetes.io/projected/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-kube-api-access-4b9hq\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434625 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434794 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-config-data\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434865 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/defcd96a-640e-44a5-881e-b45711312d7e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434905 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.434937 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-sb\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.444450 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.451016 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/defcd96a-640e-44a5-881e-b45711312d7e-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.451116 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.451858 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.453084 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-config-data\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.461894 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.464079 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-config-data\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.464996 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-ceph\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.474761 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcs8c\" (UniqueName: \"kubernetes.io/projected/defcd96a-640e-44a5-881e-b45711312d7e-kube-api-access-fcs8c\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.476581 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.478832 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-scripts\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.497215 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-scripts\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.507380 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4b9hq\" (UniqueName: \"kubernetes.io/projected/590f125d-d9fc-4847-a4b8-6f5619a6fbcd-kube-api-access-4b9hq\") pod \"manila-share-share1-0\" (UID: \"590f125d-d9fc-4847-a4b8-6f5619a6fbcd\") " pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.507817 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/defcd96a-640e-44a5-881e-b45711312d7e-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"defcd96a-640e-44a5-881e-b45711312d7e\") " pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.544094 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-sb\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.544246 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-config\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.544377 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtxz4\" (UniqueName: \"kubernetes.io/projected/1e98e71b-d235-4b50-8145-a445b8b5db8c-kube-api-access-gtxz4\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.544431 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-dns-svc\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.544495 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-nb\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.546947 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-sb\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.569101 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-config\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.570094 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-dns-svc\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.575075 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-nb\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.579964 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.604072 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.624729 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.625936 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtxz4\" (UniqueName: \"kubernetes.io/projected/1e98e71b-d235-4b50-8145-a445b8b5db8c-kube-api-access-gtxz4\") pod \"dnsmasq-dns-5cfb6864c7-7wk4c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.627087 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.639220 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.675491 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.748354 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.750289 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcdm5\" (UniqueName: \"kubernetes.io/projected/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-kube-api-access-wcdm5\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.750377 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-etc-machine-id\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.750563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-scripts\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.750659 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-logs\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.750729 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.750959 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-config-data\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.751025 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-config-data-custom\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.852933 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcdm5\" (UniqueName: \"kubernetes.io/projected/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-kube-api-access-wcdm5\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.853214 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-etc-machine-id\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.853330 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-scripts\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.853373 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-logs\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.853401 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.853503 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-config-data\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.853543 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-config-data-custom\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.856631 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-etc-machine-id\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.861843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-logs\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.862287 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.862723 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-scripts\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.864723 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-config-data-custom\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.883502 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-config-data\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:12 crc kubenswrapper[4979]: I1204 13:27:12.897954 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcdm5\" (UniqueName: \"kubernetes.io/projected/bbb3740a-8c4b-4c8b-9248-552fcb198bf7-kube-api-access-wcdm5\") pod \"manila-api-0\" (UID: \"bbb3740a-8c4b-4c8b-9248-552fcb198bf7\") " pod="openstack/manila-api-0" Dec 04 13:27:13 crc kubenswrapper[4979]: I1204 13:27:13.088897 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 04 13:27:13 crc kubenswrapper[4979]: I1204 13:27:13.236289 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 04 13:27:13 crc kubenswrapper[4979]: W1204 13:27:13.245009 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddefcd96a_640e_44a5_881e_b45711312d7e.slice/crio-d5aec7134b19de6aa76bc9c0ade1d56b5d3c08e08d5d1ecef3fc2a4abc968353 WatchSource:0}: Error finding container d5aec7134b19de6aa76bc9c0ade1d56b5d3c08e08d5d1ecef3fc2a4abc968353: Status 404 returned error can't find the container with id d5aec7134b19de6aa76bc9c0ade1d56b5d3c08e08d5d1ecef3fc2a4abc968353 Dec 04 13:27:13 crc kubenswrapper[4979]: I1204 13:27:13.457740 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 04 13:27:13 crc kubenswrapper[4979]: W1204 13:27:13.463788 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod590f125d_d9fc_4847_a4b8_6f5619a6fbcd.slice/crio-05fc633f6d8ac8da2507a7b17d8a98e67560e3b8fdfd9c16740e0345e6a25b4d WatchSource:0}: Error finding container 05fc633f6d8ac8da2507a7b17d8a98e67560e3b8fdfd9c16740e0345e6a25b4d: Status 404 returned error can't find the container with id 05fc633f6d8ac8da2507a7b17d8a98e67560e3b8fdfd9c16740e0345e6a25b4d Dec 04 13:27:13 crc kubenswrapper[4979]: W1204 13:27:13.496445 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e98e71b_d235_4b50_8145_a445b8b5db8c.slice/crio-b32e8bd4d4b8d44bea6481977ebd1adcf10fa4d8ca0ec01d8b3c10983d67893f WatchSource:0}: Error finding container b32e8bd4d4b8d44bea6481977ebd1adcf10fa4d8ca0ec01d8b3c10983d67893f: Status 404 returned error can't find the container with id b32e8bd4d4b8d44bea6481977ebd1adcf10fa4d8ca0ec01d8b3c10983d67893f Dec 04 13:27:13 crc kubenswrapper[4979]: I1204 13:27:13.498448 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cfb6864c7-7wk4c"] Dec 04 13:27:13 crc kubenswrapper[4979]: I1204 13:27:13.713183 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 04 13:27:13 crc kubenswrapper[4979]: W1204 13:27:13.713500 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbbb3740a_8c4b_4c8b_9248_552fcb198bf7.slice/crio-b3010bad71a24b0d01261123e7c7689bd13e7469eaeb3fd8f33aa9d8776f6ee1 WatchSource:0}: Error finding container b3010bad71a24b0d01261123e7c7689bd13e7469eaeb3fd8f33aa9d8776f6ee1: Status 404 returned error can't find the container with id b3010bad71a24b0d01261123e7c7689bd13e7469eaeb3fd8f33aa9d8776f6ee1 Dec 04 13:27:14 crc kubenswrapper[4979]: I1204 13:27:14.017967 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"defcd96a-640e-44a5-881e-b45711312d7e","Type":"ContainerStarted","Data":"d5aec7134b19de6aa76bc9c0ade1d56b5d3c08e08d5d1ecef3fc2a4abc968353"} Dec 04 13:27:14 crc kubenswrapper[4979]: I1204 13:27:14.020619 4979 generic.go:334] "Generic (PLEG): container finished" podID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerID="9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce" exitCode=0 Dec 04 13:27:14 crc kubenswrapper[4979]: I1204 13:27:14.020674 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" event={"ID":"1e98e71b-d235-4b50-8145-a445b8b5db8c","Type":"ContainerDied","Data":"9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce"} Dec 04 13:27:14 crc kubenswrapper[4979]: I1204 13:27:14.020697 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" event={"ID":"1e98e71b-d235-4b50-8145-a445b8b5db8c","Type":"ContainerStarted","Data":"b32e8bd4d4b8d44bea6481977ebd1adcf10fa4d8ca0ec01d8b3c10983d67893f"} Dec 04 13:27:14 crc kubenswrapper[4979]: I1204 13:27:14.022846 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"bbb3740a-8c4b-4c8b-9248-552fcb198bf7","Type":"ContainerStarted","Data":"b3010bad71a24b0d01261123e7c7689bd13e7469eaeb3fd8f33aa9d8776f6ee1"} Dec 04 13:27:14 crc kubenswrapper[4979]: I1204 13:27:14.024739 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"590f125d-d9fc-4847-a4b8-6f5619a6fbcd","Type":"ContainerStarted","Data":"05fc633f6d8ac8da2507a7b17d8a98e67560e3b8fdfd9c16740e0345e6a25b4d"} Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.038684 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"defcd96a-640e-44a5-881e-b45711312d7e","Type":"ContainerStarted","Data":"699384be75349f607a9792b51f4127224d87048ae572c5af24edd7793d9f6876"} Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.039073 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"defcd96a-640e-44a5-881e-b45711312d7e","Type":"ContainerStarted","Data":"fdfcbc77b140cf71ee452da45e6d92a48400280813df3805d6f7371289e68ae8"} Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.041433 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" event={"ID":"1e98e71b-d235-4b50-8145-a445b8b5db8c","Type":"ContainerStarted","Data":"c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45"} Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.041692 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.046393 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"bbb3740a-8c4b-4c8b-9248-552fcb198bf7","Type":"ContainerStarted","Data":"8d1375fff7e49e33c656a7983704fb81fb93b3f827f8d04151c816f9d4035c07"} Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.046432 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"bbb3740a-8c4b-4c8b-9248-552fcb198bf7","Type":"ContainerStarted","Data":"6db38730c2bcb1c7598a358ce3a7b605ec45afe24d2d5d6720119c161527a0a4"} Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.046554 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.061637 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.348396306 podStartE2EDuration="3.06161531s" podCreationTimestamp="2025-12-04 13:27:12 +0000 UTC" firstStartedPulling="2025-12-04 13:27:13.246968851 +0000 UTC m=+6257.521264655" lastFinishedPulling="2025-12-04 13:27:13.960187865 +0000 UTC m=+6258.234483659" observedRunningTime="2025-12-04 13:27:15.059772011 +0000 UTC m=+6259.334067835" watchObservedRunningTime="2025-12-04 13:27:15.06161531 +0000 UTC m=+6259.335911124" Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.091167 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.091150102 podStartE2EDuration="3.091150102s" podCreationTimestamp="2025-12-04 13:27:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:27:15.08212904 +0000 UTC m=+6259.356424854" watchObservedRunningTime="2025-12-04 13:27:15.091150102 +0000 UTC m=+6259.365445906" Dec 04 13:27:15 crc kubenswrapper[4979]: I1204 13:27:15.110276 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" podStartSLOduration=3.110259104 podStartE2EDuration="3.110259104s" podCreationTimestamp="2025-12-04 13:27:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:27:15.107631744 +0000 UTC m=+6259.381927568" watchObservedRunningTime="2025-12-04 13:27:15.110259104 +0000 UTC m=+6259.384554908" Dec 04 13:27:20 crc kubenswrapper[4979]: I1204 13:27:20.118862 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 13:27:21 crc kubenswrapper[4979]: I1204 13:27:21.137930 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"590f125d-d9fc-4847-a4b8-6f5619a6fbcd","Type":"ContainerStarted","Data":"d1c2451bb2efb38b287e395bd48b6a42e58b6b282f7da2685431f1441fa71e6b"} Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.149526 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"590f125d-d9fc-4847-a4b8-6f5619a6fbcd","Type":"ContainerStarted","Data":"e617288802166f63515bb5501df960712f16d600cc1766c1b799fa32ae57e289"} Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.177408 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.171081465 podStartE2EDuration="10.177385325s" podCreationTimestamp="2025-12-04 13:27:12 +0000 UTC" firstStartedPulling="2025-12-04 13:27:13.469652362 +0000 UTC m=+6257.743948166" lastFinishedPulling="2025-12-04 13:27:20.475956222 +0000 UTC m=+6264.750252026" observedRunningTime="2025-12-04 13:27:22.166977766 +0000 UTC m=+6266.441273560" watchObservedRunningTime="2025-12-04 13:27:22.177385325 +0000 UTC m=+6266.451681129" Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.581246 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.606074 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.751489 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.837540 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9fb795d9-q7m4c"] Dec 04 13:27:22 crc kubenswrapper[4979]: I1204 13:27:22.837774 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" containerName="dnsmasq-dns" containerID="cri-o://f040e9cda831fce6c18d75bbcf1a459850e3ec2e521cbc11db2f22899afc862f" gracePeriod=10 Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.179537 4979 generic.go:334] "Generic (PLEG): container finished" podID="3bb84134-7a17-451e-bf57-b70469ec7581" containerID="f040e9cda831fce6c18d75bbcf1a459850e3ec2e521cbc11db2f22899afc862f" exitCode=0 Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.179650 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" event={"ID":"3bb84134-7a17-451e-bf57-b70469ec7581","Type":"ContainerDied","Data":"f040e9cda831fce6c18d75bbcf1a459850e3ec2e521cbc11db2f22899afc862f"} Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.681807 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.733973 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc\") pod \"3bb84134-7a17-451e-bf57-b70469ec7581\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.734181 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-config\") pod \"3bb84134-7a17-451e-bf57-b70469ec7581\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.734233 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnwgn\" (UniqueName: \"kubernetes.io/projected/3bb84134-7a17-451e-bf57-b70469ec7581-kube-api-access-fnwgn\") pod \"3bb84134-7a17-451e-bf57-b70469ec7581\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.734275 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-nb\") pod \"3bb84134-7a17-451e-bf57-b70469ec7581\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.734322 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-sb\") pod \"3bb84134-7a17-451e-bf57-b70469ec7581\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.743505 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bb84134-7a17-451e-bf57-b70469ec7581-kube-api-access-fnwgn" (OuterVolumeSpecName: "kube-api-access-fnwgn") pod "3bb84134-7a17-451e-bf57-b70469ec7581" (UID: "3bb84134-7a17-451e-bf57-b70469ec7581"). InnerVolumeSpecName "kube-api-access-fnwgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.815460 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-config" (OuterVolumeSpecName: "config") pod "3bb84134-7a17-451e-bf57-b70469ec7581" (UID: "3bb84134-7a17-451e-bf57-b70469ec7581"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.828092 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3bb84134-7a17-451e-bf57-b70469ec7581" (UID: "3bb84134-7a17-451e-bf57-b70469ec7581"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.836193 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3bb84134-7a17-451e-bf57-b70469ec7581" (UID: "3bb84134-7a17-451e-bf57-b70469ec7581"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.836380 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc\") pod \"3bb84134-7a17-451e-bf57-b70469ec7581\" (UID: \"3bb84134-7a17-451e-bf57-b70469ec7581\") " Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.837093 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.837111 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnwgn\" (UniqueName: \"kubernetes.io/projected/3bb84134-7a17-451e-bf57-b70469ec7581-kube-api-access-fnwgn\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.837120 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:23 crc kubenswrapper[4979]: W1204 13:27:23.837190 4979 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/3bb84134-7a17-451e-bf57-b70469ec7581/volumes/kubernetes.io~configmap/dns-svc Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.837200 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3bb84134-7a17-451e-bf57-b70469ec7581" (UID: "3bb84134-7a17-451e-bf57-b70469ec7581"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.840449 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3bb84134-7a17-451e-bf57-b70469ec7581" (UID: "3bb84134-7a17-451e-bf57-b70469ec7581"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.940207 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:23 crc kubenswrapper[4979]: I1204 13:27:23.940248 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3bb84134-7a17-451e-bf57-b70469ec7581-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:24 crc kubenswrapper[4979]: I1204 13:27:24.194616 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" event={"ID":"3bb84134-7a17-451e-bf57-b70469ec7581","Type":"ContainerDied","Data":"814050b30a584ca25b99647b5daed0e88dc231f6c973012262b51e48fc9a2b71"} Dec 04 13:27:24 crc kubenswrapper[4979]: I1204 13:27:24.194970 4979 scope.go:117] "RemoveContainer" containerID="f040e9cda831fce6c18d75bbcf1a459850e3ec2e521cbc11db2f22899afc862f" Dec 04 13:27:24 crc kubenswrapper[4979]: I1204 13:27:24.194682 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9fb795d9-q7m4c" Dec 04 13:27:24 crc kubenswrapper[4979]: I1204 13:27:24.226395 4979 scope.go:117] "RemoveContainer" containerID="dd404aafd97ae5cca6a005f48098e545faada82d16b9c4b5cd05606914e07047" Dec 04 13:27:24 crc kubenswrapper[4979]: I1204 13:27:24.249433 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9fb795d9-q7m4c"] Dec 04 13:27:24 crc kubenswrapper[4979]: I1204 13:27:24.257596 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9fb795d9-q7m4c"] Dec 04 13:27:25 crc kubenswrapper[4979]: I1204 13:27:25.575969 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:27:25 crc kubenswrapper[4979]: I1204 13:27:25.576735 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-central-agent" containerID="cri-o://2d2e7536d779e8b98085c87e5c3009706b53d62e29334f38404154b233f63f23" gracePeriod=30 Dec 04 13:27:25 crc kubenswrapper[4979]: I1204 13:27:25.576803 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="sg-core" containerID="cri-o://ad156b62041007ce7aef25003b37b0fc7768765fbae7d6e6d427c0bb29695aa6" gracePeriod=30 Dec 04 13:27:25 crc kubenswrapper[4979]: I1204 13:27:25.576864 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-notification-agent" containerID="cri-o://ae8391deaa4546d7f0ccc9c955c02efe5e501c91c6d7359d75693c84a5b6c8b9" gracePeriod=30 Dec 04 13:27:25 crc kubenswrapper[4979]: I1204 13:27:25.576953 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="proxy-httpd" containerID="cri-o://c60543d267b30eb046a7935f6576af57b1774f574a5c7cd49274e51f8e8144e1" gracePeriod=30 Dec 04 13:27:26 crc kubenswrapper[4979]: I1204 13:27:26.218900 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" path="/var/lib/kubelet/pods/3bb84134-7a17-451e-bf57-b70469ec7581/volumes" Dec 04 13:27:26 crc kubenswrapper[4979]: I1204 13:27:26.229506 4979 generic.go:334] "Generic (PLEG): container finished" podID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerID="c60543d267b30eb046a7935f6576af57b1774f574a5c7cd49274e51f8e8144e1" exitCode=0 Dec 04 13:27:26 crc kubenswrapper[4979]: I1204 13:27:26.229550 4979 generic.go:334] "Generic (PLEG): container finished" podID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerID="ad156b62041007ce7aef25003b37b0fc7768765fbae7d6e6d427c0bb29695aa6" exitCode=2 Dec 04 13:27:26 crc kubenswrapper[4979]: I1204 13:27:26.229594 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerDied","Data":"c60543d267b30eb046a7935f6576af57b1774f574a5c7cd49274e51f8e8144e1"} Dec 04 13:27:26 crc kubenswrapper[4979]: I1204 13:27:26.229644 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerDied","Data":"ad156b62041007ce7aef25003b37b0fc7768765fbae7d6e6d427c0bb29695aa6"} Dec 04 13:27:27 crc kubenswrapper[4979]: I1204 13:27:27.242325 4979 generic.go:334] "Generic (PLEG): container finished" podID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerID="2d2e7536d779e8b98085c87e5c3009706b53d62e29334f38404154b233f63f23" exitCode=0 Dec 04 13:27:27 crc kubenswrapper[4979]: I1204 13:27:27.242407 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerDied","Data":"2d2e7536d779e8b98085c87e5c3009706b53d62e29334f38404154b233f63f23"} Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.307884 4979 generic.go:334] "Generic (PLEG): container finished" podID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerID="ae8391deaa4546d7f0ccc9c955c02efe5e501c91c6d7359d75693c84a5b6c8b9" exitCode=0 Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.307973 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerDied","Data":"ae8391deaa4546d7f0ccc9c955c02efe5e501c91c6d7359d75693c84a5b6c8b9"} Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.637154 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699437 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-scripts\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699586 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-log-httpd\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699631 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-sg-core-conf-yaml\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699741 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9r8k7\" (UniqueName: \"kubernetes.io/projected/15ef03c7-1056-4423-bb89-1ea324bbd3be-kube-api-access-9r8k7\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699789 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-config-data\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699811 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-run-httpd\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.699849 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-combined-ca-bundle\") pod \"15ef03c7-1056-4423-bb89-1ea324bbd3be\" (UID: \"15ef03c7-1056-4423-bb89-1ea324bbd3be\") " Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.700318 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.700518 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.700915 4979 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.700931 4979 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15ef03c7-1056-4423-bb89-1ea324bbd3be-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.706655 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15ef03c7-1056-4423-bb89-1ea324bbd3be-kube-api-access-9r8k7" (OuterVolumeSpecName: "kube-api-access-9r8k7") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "kube-api-access-9r8k7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.719441 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-scripts" (OuterVolumeSpecName: "scripts") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.754429 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.802962 4979 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.803002 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9r8k7\" (UniqueName: \"kubernetes.io/projected/15ef03c7-1056-4423-bb89-1ea324bbd3be-kube-api-access-9r8k7\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.803015 4979 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-scripts\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.804156 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.850571 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-config-data" (OuterVolumeSpecName: "config-data") pod "15ef03c7-1056-4423-bb89-1ea324bbd3be" (UID: "15ef03c7-1056-4423-bb89-1ea324bbd3be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.905492 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:30 crc kubenswrapper[4979]: I1204 13:27:30.905557 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15ef03c7-1056-4423-bb89-1ea324bbd3be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.322828 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15ef03c7-1056-4423-bb89-1ea324bbd3be","Type":"ContainerDied","Data":"7226259a55a84cc4c582e2babf823fc407d1896558e32a53ce297d19861a3139"} Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.323254 4979 scope.go:117] "RemoveContainer" containerID="c60543d267b30eb046a7935f6576af57b1774f574a5c7cd49274e51f8e8144e1" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.322977 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.367458 4979 scope.go:117] "RemoveContainer" containerID="ad156b62041007ce7aef25003b37b0fc7768765fbae7d6e6d427c0bb29695aa6" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.374199 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.384387 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.402133 4979 scope.go:117] "RemoveContainer" containerID="ae8391deaa4546d7f0ccc9c955c02efe5e501c91c6d7359d75693c84a5b6c8b9" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414172 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:27:31 crc kubenswrapper[4979]: E1204 13:27:31.414669 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="sg-core" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414706 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="sg-core" Dec 04 13:27:31 crc kubenswrapper[4979]: E1204 13:27:31.414730 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" containerName="dnsmasq-dns" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414736 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" containerName="dnsmasq-dns" Dec 04 13:27:31 crc kubenswrapper[4979]: E1204 13:27:31.414766 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-central-agent" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414775 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-central-agent" Dec 04 13:27:31 crc kubenswrapper[4979]: E1204 13:27:31.414792 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="proxy-httpd" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414798 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="proxy-httpd" Dec 04 13:27:31 crc kubenswrapper[4979]: E1204 13:27:31.414810 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" containerName="init" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414816 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" containerName="init" Dec 04 13:27:31 crc kubenswrapper[4979]: E1204 13:27:31.414827 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-notification-agent" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.414833 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-notification-agent" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.415045 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-central-agent" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.415060 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="proxy-httpd" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.415073 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="sg-core" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.415096 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" containerName="ceilometer-notification-agent" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.415106 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bb84134-7a17-451e-bf57-b70469ec7581" containerName="dnsmasq-dns" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.417108 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.420258 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.421045 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.431946 4979 scope.go:117] "RemoveContainer" containerID="2d2e7536d779e8b98085c87e5c3009706b53d62e29334f38404154b233f63f23" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.436533 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.517714 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-log-httpd\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.517759 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.517787 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnzg4\" (UniqueName: \"kubernetes.io/projected/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-kube-api-access-vnzg4\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.517906 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-scripts\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.517924 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-config-data\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.518014 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.518066 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-run-httpd\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620458 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-log-httpd\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620511 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620550 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnzg4\" (UniqueName: \"kubernetes.io/projected/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-kube-api-access-vnzg4\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620600 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-scripts\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620626 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-config-data\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620720 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.620768 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-run-httpd\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.621280 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-run-httpd\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.621374 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-log-httpd\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.625808 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.628856 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.629378 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-scripts\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.634806 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-config-data\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.646370 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnzg4\" (UniqueName: \"kubernetes.io/projected/462e9872-5a8f-4fd0-bd81-cf2ec4471b60-kube-api-access-vnzg4\") pod \"ceilometer-0\" (UID: \"462e9872-5a8f-4fd0-bd81-cf2ec4471b60\") " pod="openstack/ceilometer-0" Dec 04 13:27:31 crc kubenswrapper[4979]: I1204 13:27:31.756857 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 04 13:27:32 crc kubenswrapper[4979]: I1204 13:27:32.214502 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15ef03c7-1056-4423-bb89-1ea324bbd3be" path="/var/lib/kubelet/pods/15ef03c7-1056-4423-bb89-1ea324bbd3be/volumes" Dec 04 13:27:32 crc kubenswrapper[4979]: I1204 13:27:32.262387 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 04 13:27:32 crc kubenswrapper[4979]: I1204 13:27:32.338058 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"462e9872-5a8f-4fd0-bd81-cf2ec4471b60","Type":"ContainerStarted","Data":"4c369379fddcb77b2452e9d49626d9e1cb1c840e7069f8a3e376d109c52ff07e"} Dec 04 13:27:33 crc kubenswrapper[4979]: I1204 13:27:33.352787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"462e9872-5a8f-4fd0-bd81-cf2ec4471b60","Type":"ContainerStarted","Data":"54a9993cd1aabc879d151ab8dd454db92c42dc1f1c4f4e06bddb994073dbafdb"} Dec 04 13:27:34 crc kubenswrapper[4979]: I1204 13:27:34.367535 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"462e9872-5a8f-4fd0-bd81-cf2ec4471b60","Type":"ContainerStarted","Data":"8b79efdbec0ade0121eeb8c2981a13d3d7b72778c73fa4729da6133b7c1117c2"} Dec 04 13:27:34 crc kubenswrapper[4979]: I1204 13:27:34.956420 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 04 13:27:34 crc kubenswrapper[4979]: I1204 13:27:34.978031 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 04 13:27:35 crc kubenswrapper[4979]: I1204 13:27:35.381460 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"462e9872-5a8f-4fd0-bd81-cf2ec4471b60","Type":"ContainerStarted","Data":"aed8eeacf4267eb51ebab425077dc97c28655c7503bb026ae97a792ac464cc7d"} Dec 04 13:27:35 crc kubenswrapper[4979]: I1204 13:27:35.431587 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 04 13:27:36 crc kubenswrapper[4979]: I1204 13:27:36.396407 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"462e9872-5a8f-4fd0-bd81-cf2ec4471b60","Type":"ContainerStarted","Data":"8070e30fc08a3d21dc243cba1212f2a1ea710928562be120545b76968f1aa601"} Dec 04 13:27:36 crc kubenswrapper[4979]: I1204 13:27:36.448961 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.7474167550000002 podStartE2EDuration="5.448931788s" podCreationTimestamp="2025-12-04 13:27:31 +0000 UTC" firstStartedPulling="2025-12-04 13:27:32.270428144 +0000 UTC m=+6276.544723948" lastFinishedPulling="2025-12-04 13:27:35.971943177 +0000 UTC m=+6280.246238981" observedRunningTime="2025-12-04 13:27:36.433603147 +0000 UTC m=+6280.707898951" watchObservedRunningTime="2025-12-04 13:27:36.448931788 +0000 UTC m=+6280.723227592" Dec 04 13:27:37 crc kubenswrapper[4979]: I1204 13:27:37.406514 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 04 13:27:43 crc kubenswrapper[4979]: I1204 13:27:43.050560 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8f96-account-create-update-pcds5"] Dec 04 13:27:43 crc kubenswrapper[4979]: I1204 13:27:43.063087 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-b4xr9"] Dec 04 13:27:43 crc kubenswrapper[4979]: I1204 13:27:43.071671 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8f96-account-create-update-pcds5"] Dec 04 13:27:43 crc kubenswrapper[4979]: I1204 13:27:43.080044 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-b4xr9"] Dec 04 13:27:44 crc kubenswrapper[4979]: I1204 13:27:44.212653 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d8f7d50-afdf-4be5-b4ce-d7cf262369f2" path="/var/lib/kubelet/pods/0d8f7d50-afdf-4be5-b4ce-d7cf262369f2/volumes" Dec 04 13:27:44 crc kubenswrapper[4979]: I1204 13:27:44.214518 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f7be7de-5236-490b-b001-a9bfbb6f4575" path="/var/lib/kubelet/pods/5f7be7de-5236-490b-b001-a9bfbb6f4575/volumes" Dec 04 13:27:51 crc kubenswrapper[4979]: I1204 13:27:51.047319 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-55dkm"] Dec 04 13:27:51 crc kubenswrapper[4979]: I1204 13:27:51.065523 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-55dkm"] Dec 04 13:27:52 crc kubenswrapper[4979]: I1204 13:27:52.211292 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31eb23cc-ca38-4fa6-926b-549b3a4eb35a" path="/var/lib/kubelet/pods/31eb23cc-ca38-4fa6-926b-549b3a4eb35a/volumes" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.484664 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mk966"] Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.491153 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.538960 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mk966"] Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.545672 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh59m\" (UniqueName: \"kubernetes.io/projected/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-kube-api-access-qh59m\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.545782 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-utilities\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.545917 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-catalog-content\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.648416 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-catalog-content\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.648563 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh59m\" (UniqueName: \"kubernetes.io/projected/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-kube-api-access-qh59m\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.648631 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-utilities\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.649206 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-utilities\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.649289 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-catalog-content\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.670687 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh59m\" (UniqueName: \"kubernetes.io/projected/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-kube-api-access-qh59m\") pod \"certified-operators-mk966\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.762530 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 04 13:28:01 crc kubenswrapper[4979]: I1204 13:28:01.833419 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:02 crc kubenswrapper[4979]: I1204 13:28:02.383809 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mk966"] Dec 04 13:28:02 crc kubenswrapper[4979]: I1204 13:28:02.701267 4979 generic.go:334] "Generic (PLEG): container finished" podID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerID="417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f" exitCode=0 Dec 04 13:28:02 crc kubenswrapper[4979]: I1204 13:28:02.701451 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerDied","Data":"417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f"} Dec 04 13:28:02 crc kubenswrapper[4979]: I1204 13:28:02.702529 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerStarted","Data":"8c992dcdb7ff4a779858d16afb5401b89ded3397bbc744d9819b53a1349dbd7e"} Dec 04 13:28:03 crc kubenswrapper[4979]: I1204 13:28:03.714808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerStarted","Data":"70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380"} Dec 04 13:28:05 crc kubenswrapper[4979]: I1204 13:28:05.735422 4979 generic.go:334] "Generic (PLEG): container finished" podID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerID="70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380" exitCode=0 Dec 04 13:28:05 crc kubenswrapper[4979]: I1204 13:28:05.735484 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerDied","Data":"70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380"} Dec 04 13:28:06 crc kubenswrapper[4979]: I1204 13:28:06.750409 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerStarted","Data":"2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0"} Dec 04 13:28:06 crc kubenswrapper[4979]: I1204 13:28:06.773074 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mk966" podStartSLOduration=2.22049104 podStartE2EDuration="5.77305119s" podCreationTimestamp="2025-12-04 13:28:01 +0000 UTC" firstStartedPulling="2025-12-04 13:28:02.703476707 +0000 UTC m=+6306.977772501" lastFinishedPulling="2025-12-04 13:28:06.256036847 +0000 UTC m=+6310.530332651" observedRunningTime="2025-12-04 13:28:06.768432887 +0000 UTC m=+6311.042728721" watchObservedRunningTime="2025-12-04 13:28:06.77305119 +0000 UTC m=+6311.047346994" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.358960 4979 scope.go:117] "RemoveContainer" containerID="16a9c62335bfcd5c14617e10febb89109ba8c1f5269d781f97ab41ee40069b01" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.383672 4979 scope.go:117] "RemoveContainer" containerID="9dea7d5bc9a412bd09350b4cbeed6fdffea9b2665cfc71a3002c95e093136be8" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.449226 4979 scope.go:117] "RemoveContainer" containerID="083d4f5c7da426c8af9e3254c77469de6f04118b249b0abef98f11b3d015a153" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.490969 4979 scope.go:117] "RemoveContainer" containerID="f95b5dff05ff1aa32bd72dd2810c3ca8678f54b7f48acace75412084cb40fe5a" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.546729 4979 scope.go:117] "RemoveContainer" containerID="761b4d1abe60245e5b98dc10b689c438d24fce94b1a84dcdcce6a7ecb3d24cd5" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.592701 4979 scope.go:117] "RemoveContainer" containerID="77be6ee27918f9b9d877e3fd44caeb0aec67223551070665f7aa95b9636e1b7d" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.834583 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.834634 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:11 crc kubenswrapper[4979]: I1204 13:28:11.882989 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:12 crc kubenswrapper[4979]: I1204 13:28:12.878543 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:12 crc kubenswrapper[4979]: I1204 13:28:12.930785 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mk966"] Dec 04 13:28:14 crc kubenswrapper[4979]: I1204 13:28:14.841942 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mk966" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="registry-server" containerID="cri-o://2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0" gracePeriod=2 Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.423286 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.439533 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-catalog-content\") pod \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.439653 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh59m\" (UniqueName: \"kubernetes.io/projected/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-kube-api-access-qh59m\") pod \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.439748 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-utilities\") pod \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\" (UID: \"b6b31358-efaf-4801-a93b-c0aebbc5bf4b\") " Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.441045 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-utilities" (OuterVolumeSpecName: "utilities") pod "b6b31358-efaf-4801-a93b-c0aebbc5bf4b" (UID: "b6b31358-efaf-4801-a93b-c0aebbc5bf4b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.451437 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-kube-api-access-qh59m" (OuterVolumeSpecName: "kube-api-access-qh59m") pod "b6b31358-efaf-4801-a93b-c0aebbc5bf4b" (UID: "b6b31358-efaf-4801-a93b-c0aebbc5bf4b"). InnerVolumeSpecName "kube-api-access-qh59m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.521854 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b6b31358-efaf-4801-a93b-c0aebbc5bf4b" (UID: "b6b31358-efaf-4801-a93b-c0aebbc5bf4b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.542023 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.542103 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh59m\" (UniqueName: \"kubernetes.io/projected/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-kube-api-access-qh59m\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.542115 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b6b31358-efaf-4801-a93b-c0aebbc5bf4b-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.854731 4979 generic.go:334] "Generic (PLEG): container finished" podID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerID="2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0" exitCode=0 Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.854791 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerDied","Data":"2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0"} Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.854831 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mk966" event={"ID":"b6b31358-efaf-4801-a93b-c0aebbc5bf4b","Type":"ContainerDied","Data":"8c992dcdb7ff4a779858d16afb5401b89ded3397bbc744d9819b53a1349dbd7e"} Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.854856 4979 scope.go:117] "RemoveContainer" containerID="2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.854792 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mk966" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.886104 4979 scope.go:117] "RemoveContainer" containerID="70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.894381 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mk966"] Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.906446 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mk966"] Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.930752 4979 scope.go:117] "RemoveContainer" containerID="417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.968240 4979 scope.go:117] "RemoveContainer" containerID="2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0" Dec 04 13:28:15 crc kubenswrapper[4979]: E1204 13:28:15.968727 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0\": container with ID starting with 2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0 not found: ID does not exist" containerID="2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.968773 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0"} err="failed to get container status \"2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0\": rpc error: code = NotFound desc = could not find container \"2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0\": container with ID starting with 2db419d06afce8ea0a5b5658c9e7672a6551f9581ef9ac4b3e19e16c11a1d9b0 not found: ID does not exist" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.968804 4979 scope.go:117] "RemoveContainer" containerID="70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380" Dec 04 13:28:15 crc kubenswrapper[4979]: E1204 13:28:15.969035 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380\": container with ID starting with 70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380 not found: ID does not exist" containerID="70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.969061 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380"} err="failed to get container status \"70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380\": rpc error: code = NotFound desc = could not find container \"70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380\": container with ID starting with 70612bdebafe167d88989cee32ca50b4d26658199db358a804d79e30e87f9380 not found: ID does not exist" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.969075 4979 scope.go:117] "RemoveContainer" containerID="417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f" Dec 04 13:28:15 crc kubenswrapper[4979]: E1204 13:28:15.969289 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f\": container with ID starting with 417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f not found: ID does not exist" containerID="417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f" Dec 04 13:28:15 crc kubenswrapper[4979]: I1204 13:28:15.969334 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f"} err="failed to get container status \"417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f\": rpc error: code = NotFound desc = could not find container \"417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f\": container with ID starting with 417123ef61bb3e80f17ff20bce65eefa4714349d8f9808130cfd6db0efa6ca4f not found: ID does not exist" Dec 04 13:28:16 crc kubenswrapper[4979]: I1204 13:28:16.236338 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" path="/var/lib/kubelet/pods/b6b31358-efaf-4801-a93b-c0aebbc5bf4b/volumes" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.982719 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-df7c6669f-8vdc2"] Dec 04 13:28:23 crc kubenswrapper[4979]: E1204 13:28:23.984860 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="extract-content" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.984969 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="extract-content" Dec 04 13:28:23 crc kubenswrapper[4979]: E1204 13:28:23.985043 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="registry-server" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.985098 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="registry-server" Dec 04 13:28:23 crc kubenswrapper[4979]: E1204 13:28:23.985173 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="extract-utilities" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.985232 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="extract-utilities" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.985516 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6b31358-efaf-4801-a93b-c0aebbc5bf4b" containerName="registry-server" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.986825 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.990940 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1" Dec 04 13:28:23 crc kubenswrapper[4979]: I1204 13:28:23.998476 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-df7c6669f-8vdc2"] Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.122875 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g465d\" (UniqueName: \"kubernetes.io/projected/a7f1fc01-e038-4ae2-b718-e85b6b46a106-kube-api-access-g465d\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.122999 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-dns-svc\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.123189 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-nb\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.123570 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-config\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.123636 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-sb\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.123680 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-openstack-cell1\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.226138 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-config\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.226187 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-sb\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.226219 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-openstack-cell1\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.226325 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g465d\" (UniqueName: \"kubernetes.io/projected/a7f1fc01-e038-4ae2-b718-e85b6b46a106-kube-api-access-g465d\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.226413 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-dns-svc\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.226501 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-nb\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.227106 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-config\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.227549 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-openstack-cell1\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.227621 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-sb\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.227702 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-dns-svc\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.227745 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-nb\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.248087 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g465d\" (UniqueName: \"kubernetes.io/projected/a7f1fc01-e038-4ae2-b718-e85b6b46a106-kube-api-access-g465d\") pod \"dnsmasq-dns-df7c6669f-8vdc2\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.321629 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.826948 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-df7c6669f-8vdc2"] Dec 04 13:28:24 crc kubenswrapper[4979]: I1204 13:28:24.965079 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" event={"ID":"a7f1fc01-e038-4ae2-b718-e85b6b46a106","Type":"ContainerStarted","Data":"85ec105f278ee6db2a8d3764d0a4a7ee8c62e5826e1f2b70c6786f2f4b1519c0"} Dec 04 13:28:25 crc kubenswrapper[4979]: I1204 13:28:25.977788 4979 generic.go:334] "Generic (PLEG): container finished" podID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerID="1273624e27b72807a002d9e0b7ce95e34b44df281e4ee7f8b361a5ae2827d23e" exitCode=0 Dec 04 13:28:25 crc kubenswrapper[4979]: I1204 13:28:25.977863 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" event={"ID":"a7f1fc01-e038-4ae2-b718-e85b6b46a106","Type":"ContainerDied","Data":"1273624e27b72807a002d9e0b7ce95e34b44df281e4ee7f8b361a5ae2827d23e"} Dec 04 13:28:26 crc kubenswrapper[4979]: I1204 13:28:26.990852 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" event={"ID":"a7f1fc01-e038-4ae2-b718-e85b6b46a106","Type":"ContainerStarted","Data":"b69b32ecd5ba941e205f9e7fb6240045ff10237a77b305bde4378eff21002ca9"} Dec 04 13:28:26 crc kubenswrapper[4979]: I1204 13:28:26.991493 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:27 crc kubenswrapper[4979]: I1204 13:28:27.015971 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" podStartSLOduration=4.015953052 podStartE2EDuration="4.015953052s" podCreationTimestamp="2025-12-04 13:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:28:27.008169563 +0000 UTC m=+6331.282465367" watchObservedRunningTime="2025-12-04 13:28:27.015953052 +0000 UTC m=+6331.290248856" Dec 04 13:28:28 crc kubenswrapper[4979]: I1204 13:28:28.040634 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:28:28 crc kubenswrapper[4979]: I1204 13:28:28.040693 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.323471 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.387199 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cfb6864c7-7wk4c"] Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.387450 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerName="dnsmasq-dns" containerID="cri-o://c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45" gracePeriod=10 Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.575221 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5498899cc5-dshwr"] Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.577930 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.594762 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5498899cc5-dshwr"] Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.655278 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-config\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.655348 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-dns-svc\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.655397 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-ovsdbserver-sb\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.655608 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-ovsdbserver-nb\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.655815 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt8sm\" (UniqueName: \"kubernetes.io/projected/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-kube-api-access-mt8sm\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.656003 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-openstack-cell1\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.757884 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-config\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.757927 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-dns-svc\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.757960 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-ovsdbserver-sb\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.758011 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-ovsdbserver-nb\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.758062 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt8sm\" (UniqueName: \"kubernetes.io/projected/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-kube-api-access-mt8sm\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.758091 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-openstack-cell1\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.759089 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-openstack-cell1\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.759617 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-config\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.760179 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-dns-svc\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.760788 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-ovsdbserver-sb\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.761511 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-ovsdbserver-nb\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.801589 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt8sm\" (UniqueName: \"kubernetes.io/projected/7574d1f6-997e-4d18-8d8a-c49ba3b0f12e-kube-api-access-mt8sm\") pod \"dnsmasq-dns-5498899cc5-dshwr\" (UID: \"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e\") " pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:34 crc kubenswrapper[4979]: I1204 13:28:34.905793 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.016268 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.085514 4979 generic.go:334] "Generic (PLEG): container finished" podID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerID="c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45" exitCode=0 Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.085556 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" event={"ID":"1e98e71b-d235-4b50-8145-a445b8b5db8c","Type":"ContainerDied","Data":"c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45"} Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.085602 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" event={"ID":"1e98e71b-d235-4b50-8145-a445b8b5db8c","Type":"ContainerDied","Data":"b32e8bd4d4b8d44bea6481977ebd1adcf10fa4d8ca0ec01d8b3c10983d67893f"} Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.085622 4979 scope.go:117] "RemoveContainer" containerID="c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.085809 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cfb6864c7-7wk4c" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.138119 4979 scope.go:117] "RemoveContainer" containerID="9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.167540 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-config\") pod \"1e98e71b-d235-4b50-8145-a445b8b5db8c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.167873 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-nb\") pod \"1e98e71b-d235-4b50-8145-a445b8b5db8c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.167964 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-sb\") pod \"1e98e71b-d235-4b50-8145-a445b8b5db8c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.168047 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-dns-svc\") pod \"1e98e71b-d235-4b50-8145-a445b8b5db8c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.168196 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtxz4\" (UniqueName: \"kubernetes.io/projected/1e98e71b-d235-4b50-8145-a445b8b5db8c-kube-api-access-gtxz4\") pod \"1e98e71b-d235-4b50-8145-a445b8b5db8c\" (UID: \"1e98e71b-d235-4b50-8145-a445b8b5db8c\") " Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.170626 4979 scope.go:117] "RemoveContainer" containerID="c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45" Dec 04 13:28:35 crc kubenswrapper[4979]: E1204 13:28:35.172975 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45\": container with ID starting with c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45 not found: ID does not exist" containerID="c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.173007 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45"} err="failed to get container status \"c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45\": rpc error: code = NotFound desc = could not find container \"c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45\": container with ID starting with c29b8716b3523ca09e2638a00dafbf107dc64d1e5c2e3c2d136b28991bc90f45 not found: ID does not exist" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.173031 4979 scope.go:117] "RemoveContainer" containerID="9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.173452 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e98e71b-d235-4b50-8145-a445b8b5db8c-kube-api-access-gtxz4" (OuterVolumeSpecName: "kube-api-access-gtxz4") pod "1e98e71b-d235-4b50-8145-a445b8b5db8c" (UID: "1e98e71b-d235-4b50-8145-a445b8b5db8c"). InnerVolumeSpecName "kube-api-access-gtxz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:28:35 crc kubenswrapper[4979]: E1204 13:28:35.187294 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce\": container with ID starting with 9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce not found: ID does not exist" containerID="9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.187350 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce"} err="failed to get container status \"9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce\": rpc error: code = NotFound desc = could not find container \"9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce\": container with ID starting with 9a939b335f77e8d5c4a746943543601bf59c18c2c7eda09dc1877bf54ab076ce not found: ID does not exist" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.251120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1e98e71b-d235-4b50-8145-a445b8b5db8c" (UID: "1e98e71b-d235-4b50-8145-a445b8b5db8c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.268868 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1e98e71b-d235-4b50-8145-a445b8b5db8c" (UID: "1e98e71b-d235-4b50-8145-a445b8b5db8c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.270203 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-config" (OuterVolumeSpecName: "config") pod "1e98e71b-d235-4b50-8145-a445b8b5db8c" (UID: "1e98e71b-d235-4b50-8145-a445b8b5db8c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.270944 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtxz4\" (UniqueName: \"kubernetes.io/projected/1e98e71b-d235-4b50-8145-a445b8b5db8c-kube-api-access-gtxz4\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.270969 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.270983 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.270995 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.273242 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1e98e71b-d235-4b50-8145-a445b8b5db8c" (UID: "1e98e71b-d235-4b50-8145-a445b8b5db8c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.373501 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1e98e71b-d235-4b50-8145-a445b8b5db8c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.424646 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cfb6864c7-7wk4c"] Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.436045 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cfb6864c7-7wk4c"] Dec 04 13:28:35 crc kubenswrapper[4979]: I1204 13:28:35.504167 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5498899cc5-dshwr"] Dec 04 13:28:36 crc kubenswrapper[4979]: I1204 13:28:36.101104 4979 generic.go:334] "Generic (PLEG): container finished" podID="7574d1f6-997e-4d18-8d8a-c49ba3b0f12e" containerID="8b2de336fbb55bdac5a52b69ccff4ab094808b14adae7003396292075c864e63" exitCode=0 Dec 04 13:28:36 crc kubenswrapper[4979]: I1204 13:28:36.101183 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" event={"ID":"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e","Type":"ContainerDied","Data":"8b2de336fbb55bdac5a52b69ccff4ab094808b14adae7003396292075c864e63"} Dec 04 13:28:36 crc kubenswrapper[4979]: I1204 13:28:36.101768 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" event={"ID":"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e","Type":"ContainerStarted","Data":"37449dc421aedabceb31bc6baf38b5d2b88d7f837b1ecda91f26001a351a83f8"} Dec 04 13:28:36 crc kubenswrapper[4979]: I1204 13:28:36.213176 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" path="/var/lib/kubelet/pods/1e98e71b-d235-4b50-8145-a445b8b5db8c/volumes" Dec 04 13:28:37 crc kubenswrapper[4979]: I1204 13:28:37.120775 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" event={"ID":"7574d1f6-997e-4d18-8d8a-c49ba3b0f12e","Type":"ContainerStarted","Data":"6efbe81d6d24d364f27bd7e42e9fe0521d4e055763129df10ad19778a4c3e318"} Dec 04 13:28:37 crc kubenswrapper[4979]: I1204 13:28:37.121545 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:37 crc kubenswrapper[4979]: I1204 13:28:37.150900 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" podStartSLOduration=3.150871714 podStartE2EDuration="3.150871714s" podCreationTimestamp="2025-12-04 13:28:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:28:37.139825567 +0000 UTC m=+6341.414121371" watchObservedRunningTime="2025-12-04 13:28:37.150871714 +0000 UTC m=+6341.425167518" Dec 04 13:28:44 crc kubenswrapper[4979]: I1204 13:28:44.910153 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5498899cc5-dshwr" Dec 04 13:28:44 crc kubenswrapper[4979]: I1204 13:28:44.973914 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-df7c6669f-8vdc2"] Dec 04 13:28:44 crc kubenswrapper[4979]: I1204 13:28:44.974158 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerName="dnsmasq-dns" containerID="cri-o://b69b32ecd5ba941e205f9e7fb6240045ff10237a77b305bde4378eff21002ca9" gracePeriod=10 Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.208634 4979 generic.go:334] "Generic (PLEG): container finished" podID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerID="b69b32ecd5ba941e205f9e7fb6240045ff10237a77b305bde4378eff21002ca9" exitCode=0 Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.209054 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" event={"ID":"a7f1fc01-e038-4ae2-b718-e85b6b46a106","Type":"ContainerDied","Data":"b69b32ecd5ba941e205f9e7fb6240045ff10237a77b305bde4378eff21002ca9"} Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.618414 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.708591 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-openstack-cell1\") pod \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.708656 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-dns-svc\") pod \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.708685 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g465d\" (UniqueName: \"kubernetes.io/projected/a7f1fc01-e038-4ae2-b718-e85b6b46a106-kube-api-access-g465d\") pod \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.708775 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-sb\") pod \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.708832 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-nb\") pod \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.708866 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-config\") pod \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\" (UID: \"a7f1fc01-e038-4ae2-b718-e85b6b46a106\") " Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.715038 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f1fc01-e038-4ae2-b718-e85b6b46a106-kube-api-access-g465d" (OuterVolumeSpecName: "kube-api-access-g465d") pod "a7f1fc01-e038-4ae2-b718-e85b6b46a106" (UID: "a7f1fc01-e038-4ae2-b718-e85b6b46a106"). InnerVolumeSpecName "kube-api-access-g465d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.774081 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a7f1fc01-e038-4ae2-b718-e85b6b46a106" (UID: "a7f1fc01-e038-4ae2-b718-e85b6b46a106"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.777233 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-openstack-cell1" (OuterVolumeSpecName: "openstack-cell1") pod "a7f1fc01-e038-4ae2-b718-e85b6b46a106" (UID: "a7f1fc01-e038-4ae2-b718-e85b6b46a106"). InnerVolumeSpecName "openstack-cell1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.777673 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a7f1fc01-e038-4ae2-b718-e85b6b46a106" (UID: "a7f1fc01-e038-4ae2-b718-e85b6b46a106"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.778355 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a7f1fc01-e038-4ae2-b718-e85b6b46a106" (UID: "a7f1fc01-e038-4ae2-b718-e85b6b46a106"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.782842 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-config" (OuterVolumeSpecName: "config") pod "a7f1fc01-e038-4ae2-b718-e85b6b46a106" (UID: "a7f1fc01-e038-4ae2-b718-e85b6b46a106"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.811803 4979 reconciler_common.go:293] "Volume detached for volume \"openstack-cell1\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.811847 4979 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.811890 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g465d\" (UniqueName: \"kubernetes.io/projected/a7f1fc01-e038-4ae2-b718-e85b6b46a106-kube-api-access-g465d\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.811908 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.811923 4979 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:45 crc kubenswrapper[4979]: I1204 13:28:45.811939 4979 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7f1fc01-e038-4ae2-b718-e85b6b46a106-config\") on node \"crc\" DevicePath \"\"" Dec 04 13:28:46 crc kubenswrapper[4979]: I1204 13:28:46.221090 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" event={"ID":"a7f1fc01-e038-4ae2-b718-e85b6b46a106","Type":"ContainerDied","Data":"85ec105f278ee6db2a8d3764d0a4a7ee8c62e5826e1f2b70c6786f2f4b1519c0"} Dec 04 13:28:46 crc kubenswrapper[4979]: I1204 13:28:46.221143 4979 scope.go:117] "RemoveContainer" containerID="b69b32ecd5ba941e205f9e7fb6240045ff10237a77b305bde4378eff21002ca9" Dec 04 13:28:46 crc kubenswrapper[4979]: I1204 13:28:46.221161 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-df7c6669f-8vdc2" Dec 04 13:28:46 crc kubenswrapper[4979]: I1204 13:28:46.252319 4979 scope.go:117] "RemoveContainer" containerID="1273624e27b72807a002d9e0b7ce95e34b44df281e4ee7f8b361a5ae2827d23e" Dec 04 13:28:46 crc kubenswrapper[4979]: I1204 13:28:46.255358 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-df7c6669f-8vdc2"] Dec 04 13:28:46 crc kubenswrapper[4979]: I1204 13:28:46.267108 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-df7c6669f-8vdc2"] Dec 04 13:28:48 crc kubenswrapper[4979]: I1204 13:28:48.214904 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" path="/var/lib/kubelet/pods/a7f1fc01-e038-4ae2-b718-e85b6b46a106/volumes" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.856321 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf"] Dec 04 13:28:55 crc kubenswrapper[4979]: E1204 13:28:55.857290 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerName="init" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.857327 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerName="init" Dec 04 13:28:55 crc kubenswrapper[4979]: E1204 13:28:55.857353 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerName="dnsmasq-dns" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.857359 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerName="dnsmasq-dns" Dec 04 13:28:55 crc kubenswrapper[4979]: E1204 13:28:55.857384 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerName="init" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.857392 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerName="init" Dec 04 13:28:55 crc kubenswrapper[4979]: E1204 13:28:55.857439 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerName="dnsmasq-dns" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.857446 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerName="dnsmasq-dns" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.858011 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f1fc01-e038-4ae2-b718-e85b6b46a106" containerName="dnsmasq-dns" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.858036 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e98e71b-d235-4b50-8145-a445b8b5db8c" containerName="dnsmasq-dns" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.858905 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.862680 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.863138 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.863405 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.863585 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.881387 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf"] Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.934979 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.935034 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rflkc\" (UniqueName: \"kubernetes.io/projected/39f115b5-6b82-4b63-848e-3cc6ae15099e-kube-api-access-rflkc\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.935141 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.935180 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:55 crc kubenswrapper[4979]: I1204 13:28:55.935250 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.037649 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.038020 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.038090 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.038146 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.038175 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rflkc\" (UniqueName: \"kubernetes.io/projected/39f115b5-6b82-4b63-848e-3cc6ae15099e-kube-api-access-rflkc\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.043842 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-pre-adoption-validation-combined-ca-bundle\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.048371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-inventory\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.048389 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ssh-key\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.048690 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ceph\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.055736 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rflkc\" (UniqueName: \"kubernetes.io/projected/39f115b5-6b82-4b63-848e-3cc6ae15099e-kube-api-access-rflkc\") pod \"pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.214910 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.223175 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.757472 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf"] Dec 04 13:28:56 crc kubenswrapper[4979]: W1204 13:28:56.758849 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39f115b5_6b82_4b63_848e_3cc6ae15099e.slice/crio-e0f6ba1bf215b671dcf524b2740ce18102a3b20fe7769ec816785ac9640ad9f9 WatchSource:0}: Error finding container e0f6ba1bf215b671dcf524b2740ce18102a3b20fe7769ec816785ac9640ad9f9: Status 404 returned error can't find the container with id e0f6ba1bf215b671dcf524b2740ce18102a3b20fe7769ec816785ac9640ad9f9 Dec 04 13:28:56 crc kubenswrapper[4979]: I1204 13:28:56.761592 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:28:57 crc kubenswrapper[4979]: I1204 13:28:57.333555 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" event={"ID":"39f115b5-6b82-4b63-848e-3cc6ae15099e","Type":"ContainerStarted","Data":"e0f6ba1bf215b671dcf524b2740ce18102a3b20fe7769ec816785ac9640ad9f9"} Dec 04 13:28:58 crc kubenswrapper[4979]: I1204 13:28:58.040860 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:28:58 crc kubenswrapper[4979]: I1204 13:28:58.044582 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:29:07 crc kubenswrapper[4979]: I1204 13:29:07.367235 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:29:08 crc kubenswrapper[4979]: I1204 13:29:08.441035 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" event={"ID":"39f115b5-6b82-4b63-848e-3cc6ae15099e","Type":"ContainerStarted","Data":"cf5995194a94c6b34e14fa706000fa9299b3596760a78dfd4feb170410bea4aa"} Dec 04 13:29:08 crc kubenswrapper[4979]: I1204 13:29:08.460394 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" podStartSLOduration=2.857083308 podStartE2EDuration="13.460374049s" podCreationTimestamp="2025-12-04 13:28:55 +0000 UTC" firstStartedPulling="2025-12-04 13:28:56.761358827 +0000 UTC m=+6361.035654641" lastFinishedPulling="2025-12-04 13:29:07.364649578 +0000 UTC m=+6371.638945382" observedRunningTime="2025-12-04 13:29:08.454577594 +0000 UTC m=+6372.728873408" watchObservedRunningTime="2025-12-04 13:29:08.460374049 +0000 UTC m=+6372.734669853" Dec 04 13:29:20 crc kubenswrapper[4979]: I1204 13:29:20.563967 4979 generic.go:334] "Generic (PLEG): container finished" podID="39f115b5-6b82-4b63-848e-3cc6ae15099e" containerID="cf5995194a94c6b34e14fa706000fa9299b3596760a78dfd4feb170410bea4aa" exitCode=0 Dec 04 13:29:20 crc kubenswrapper[4979]: I1204 13:29:20.564086 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" event={"ID":"39f115b5-6b82-4b63-848e-3cc6ae15099e","Type":"ContainerDied","Data":"cf5995194a94c6b34e14fa706000fa9299b3596760a78dfd4feb170410bea4aa"} Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.020136 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.126677 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rflkc\" (UniqueName: \"kubernetes.io/projected/39f115b5-6b82-4b63-848e-3cc6ae15099e-kube-api-access-rflkc\") pod \"39f115b5-6b82-4b63-848e-3cc6ae15099e\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.126738 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-pre-adoption-validation-combined-ca-bundle\") pod \"39f115b5-6b82-4b63-848e-3cc6ae15099e\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.126907 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-inventory\") pod \"39f115b5-6b82-4b63-848e-3cc6ae15099e\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.127008 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ssh-key\") pod \"39f115b5-6b82-4b63-848e-3cc6ae15099e\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.127025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ceph\") pod \"39f115b5-6b82-4b63-848e-3cc6ae15099e\" (UID: \"39f115b5-6b82-4b63-848e-3cc6ae15099e\") " Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.154574 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ceph" (OuterVolumeSpecName: "ceph") pod "39f115b5-6b82-4b63-848e-3cc6ae15099e" (UID: "39f115b5-6b82-4b63-848e-3cc6ae15099e"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.154737 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-pre-adoption-validation-combined-ca-bundle" (OuterVolumeSpecName: "pre-adoption-validation-combined-ca-bundle") pod "39f115b5-6b82-4b63-848e-3cc6ae15099e" (UID: "39f115b5-6b82-4b63-848e-3cc6ae15099e"). InnerVolumeSpecName "pre-adoption-validation-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.155103 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39f115b5-6b82-4b63-848e-3cc6ae15099e-kube-api-access-rflkc" (OuterVolumeSpecName: "kube-api-access-rflkc") pod "39f115b5-6b82-4b63-848e-3cc6ae15099e" (UID: "39f115b5-6b82-4b63-848e-3cc6ae15099e"). InnerVolumeSpecName "kube-api-access-rflkc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.160639 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-inventory" (OuterVolumeSpecName: "inventory") pod "39f115b5-6b82-4b63-848e-3cc6ae15099e" (UID: "39f115b5-6b82-4b63-848e-3cc6ae15099e"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.160951 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "39f115b5-6b82-4b63-848e-3cc6ae15099e" (UID: "39f115b5-6b82-4b63-848e-3cc6ae15099e"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.229257 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.229451 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.229536 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.229616 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rflkc\" (UniqueName: \"kubernetes.io/projected/39f115b5-6b82-4b63-848e-3cc6ae15099e-kube-api-access-rflkc\") on node \"crc\" DevicePath \"\"" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.229684 4979 reconciler_common.go:293] "Volume detached for volume \"pre-adoption-validation-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39f115b5-6b82-4b63-848e-3cc6ae15099e-pre-adoption-validation-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.583938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" event={"ID":"39f115b5-6b82-4b63-848e-3cc6ae15099e","Type":"ContainerDied","Data":"e0f6ba1bf215b671dcf524b2740ce18102a3b20fe7769ec816785ac9640ad9f9"} Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.583992 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf" Dec 04 13:29:22 crc kubenswrapper[4979]: I1204 13:29:22.584011 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0f6ba1bf215b671dcf524b2740ce18102a3b20fe7769ec816785ac9640ad9f9" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.040703 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.041219 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.041258 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.042009 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.042063 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" gracePeriod=600 Dec 04 13:29:28 crc kubenswrapper[4979]: E1204 13:29:28.200824 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.627494 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7"] Dec 04 13:29:28 crc kubenswrapper[4979]: E1204 13:29:28.628201 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39f115b5-6b82-4b63-848e-3cc6ae15099e" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.628219 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="39f115b5-6b82-4b63-848e-3cc6ae15099e" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.628446 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="39f115b5-6b82-4b63-848e-3cc6ae15099e" containerName="pre-adoption-validation-openstack-pre-adoption-openstack-cell1" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.629159 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.634083 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.634362 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.634463 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.634501 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.649050 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7"] Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.654508 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" exitCode=0 Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.654544 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7"} Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.654577 4979 scope.go:117] "RemoveContainer" containerID="be34e428814c510aac7c63fa5d7dc77256bd5b17bcd712eed8e9b57775e43b6a" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.655325 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:29:28 crc kubenswrapper[4979]: E1204 13:29:28.655749 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.676427 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.676492 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.676534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.676639 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dbnv\" (UniqueName: \"kubernetes.io/projected/717bd66f-9118-4b81-927c-27f66aa6483a-kube-api-access-2dbnv\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.676679 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.779060 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.779106 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.779133 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.779204 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dbnv\" (UniqueName: \"kubernetes.io/projected/717bd66f-9118-4b81-927c-27f66aa6483a-kube-api-access-2dbnv\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.779244 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.785859 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-inventory\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.786015 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ceph\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.786410 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ssh-key\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.792519 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-tripleo-cleanup-combined-ca-bundle\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.797412 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dbnv\" (UniqueName: \"kubernetes.io/projected/717bd66f-9118-4b81-927c-27f66aa6483a-kube-api-access-2dbnv\") pod \"tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:28 crc kubenswrapper[4979]: I1204 13:29:28.951046 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:29:29 crc kubenswrapper[4979]: I1204 13:29:29.516057 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7"] Dec 04 13:29:29 crc kubenswrapper[4979]: I1204 13:29:29.665853 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" event={"ID":"717bd66f-9118-4b81-927c-27f66aa6483a","Type":"ContainerStarted","Data":"280645de966a386ce45e465834df8c048b34f4d8c169dd7aea6d63d08708cf72"} Dec 04 13:29:30 crc kubenswrapper[4979]: I1204 13:29:30.691562 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" event={"ID":"717bd66f-9118-4b81-927c-27f66aa6483a","Type":"ContainerStarted","Data":"b8ee396e67fd425c0385bcfc89c8fd5029a4e7a17d3b9056fb54518252de39df"} Dec 04 13:29:30 crc kubenswrapper[4979]: I1204 13:29:30.710151 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" podStartSLOduration=2.303147269 podStartE2EDuration="2.710132303s" podCreationTimestamp="2025-12-04 13:29:28 +0000 UTC" firstStartedPulling="2025-12-04 13:29:29.518768937 +0000 UTC m=+6393.793064741" lastFinishedPulling="2025-12-04 13:29:29.925753971 +0000 UTC m=+6394.200049775" observedRunningTime="2025-12-04 13:29:30.707646856 +0000 UTC m=+6394.981942680" watchObservedRunningTime="2025-12-04 13:29:30.710132303 +0000 UTC m=+6394.984428107" Dec 04 13:29:41 crc kubenswrapper[4979]: I1204 13:29:41.198913 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:29:41 crc kubenswrapper[4979]: E1204 13:29:41.199824 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:29:53 crc kubenswrapper[4979]: I1204 13:29:53.199850 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:29:53 crc kubenswrapper[4979]: E1204 13:29:53.200772 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.175786 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl"] Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.179378 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.181933 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.185581 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.195860 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl"] Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.336992 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d314ef80-82e1-4a37-9e06-340d1ca94e89-secret-volume\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.337066 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n69tf\" (UniqueName: \"kubernetes.io/projected/d314ef80-82e1-4a37-9e06-340d1ca94e89-kube-api-access-n69tf\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.337145 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d314ef80-82e1-4a37-9e06-340d1ca94e89-config-volume\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.439401 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d314ef80-82e1-4a37-9e06-340d1ca94e89-secret-volume\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.439452 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n69tf\" (UniqueName: \"kubernetes.io/projected/d314ef80-82e1-4a37-9e06-340d1ca94e89-kube-api-access-n69tf\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.439505 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d314ef80-82e1-4a37-9e06-340d1ca94e89-config-volume\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.443959 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d314ef80-82e1-4a37-9e06-340d1ca94e89-config-volume\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.451892 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d314ef80-82e1-4a37-9e06-340d1ca94e89-secret-volume\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.466533 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n69tf\" (UniqueName: \"kubernetes.io/projected/d314ef80-82e1-4a37-9e06-340d1ca94e89-kube-api-access-n69tf\") pod \"collect-profiles-29414250-t96pl\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.510486 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:00 crc kubenswrapper[4979]: I1204 13:30:00.977174 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl"] Dec 04 13:30:01 crc kubenswrapper[4979]: I1204 13:30:01.993070 4979 generic.go:334] "Generic (PLEG): container finished" podID="d314ef80-82e1-4a37-9e06-340d1ca94e89" containerID="d7a4d12effba1fd70f81e3515f84aa4a67bbc5a1e74fe0bcda4e2ae9977e0752" exitCode=0 Dec 04 13:30:01 crc kubenswrapper[4979]: I1204 13:30:01.993158 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" event={"ID":"d314ef80-82e1-4a37-9e06-340d1ca94e89","Type":"ContainerDied","Data":"d7a4d12effba1fd70f81e3515f84aa4a67bbc5a1e74fe0bcda4e2ae9977e0752"} Dec 04 13:30:01 crc kubenswrapper[4979]: I1204 13:30:01.993620 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" event={"ID":"d314ef80-82e1-4a37-9e06-340d1ca94e89","Type":"ContainerStarted","Data":"45d466a938aee79460dffbc669fa4d4936d4277ad1822e91ac28b957c4505942"} Dec 04 13:30:02 crc kubenswrapper[4979]: E1204 13:30:02.048720 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd314ef80_82e1_4a37_9e06_340d1ca94e89.slice/crio-d7a4d12effba1fd70f81e3515f84aa4a67bbc5a1e74fe0bcda4e2ae9977e0752.scope\": RecentStats: unable to find data in memory cache]" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.612336 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.709934 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n69tf\" (UniqueName: \"kubernetes.io/projected/d314ef80-82e1-4a37-9e06-340d1ca94e89-kube-api-access-n69tf\") pod \"d314ef80-82e1-4a37-9e06-340d1ca94e89\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.709993 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d314ef80-82e1-4a37-9e06-340d1ca94e89-secret-volume\") pod \"d314ef80-82e1-4a37-9e06-340d1ca94e89\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.710220 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d314ef80-82e1-4a37-9e06-340d1ca94e89-config-volume\") pod \"d314ef80-82e1-4a37-9e06-340d1ca94e89\" (UID: \"d314ef80-82e1-4a37-9e06-340d1ca94e89\") " Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.710731 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d314ef80-82e1-4a37-9e06-340d1ca94e89-config-volume" (OuterVolumeSpecName: "config-volume") pod "d314ef80-82e1-4a37-9e06-340d1ca94e89" (UID: "d314ef80-82e1-4a37-9e06-340d1ca94e89"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.716389 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d314ef80-82e1-4a37-9e06-340d1ca94e89-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d314ef80-82e1-4a37-9e06-340d1ca94e89" (UID: "d314ef80-82e1-4a37-9e06-340d1ca94e89"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.717149 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d314ef80-82e1-4a37-9e06-340d1ca94e89-kube-api-access-n69tf" (OuterVolumeSpecName: "kube-api-access-n69tf") pod "d314ef80-82e1-4a37-9e06-340d1ca94e89" (UID: "d314ef80-82e1-4a37-9e06-340d1ca94e89"). InnerVolumeSpecName "kube-api-access-n69tf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.812695 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d314ef80-82e1-4a37-9e06-340d1ca94e89-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.812989 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n69tf\" (UniqueName: \"kubernetes.io/projected/d314ef80-82e1-4a37-9e06-340d1ca94e89-kube-api-access-n69tf\") on node \"crc\" DevicePath \"\"" Dec 04 13:30:03 crc kubenswrapper[4979]: I1204 13:30:03.813003 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d314ef80-82e1-4a37-9e06-340d1ca94e89-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:30:04 crc kubenswrapper[4979]: I1204 13:30:04.045047 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" event={"ID":"d314ef80-82e1-4a37-9e06-340d1ca94e89","Type":"ContainerDied","Data":"45d466a938aee79460dffbc669fa4d4936d4277ad1822e91ac28b957c4505942"} Dec 04 13:30:04 crc kubenswrapper[4979]: I1204 13:30:04.045439 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="45d466a938aee79460dffbc669fa4d4936d4277ad1822e91ac28b957c4505942" Dec 04 13:30:04 crc kubenswrapper[4979]: I1204 13:30:04.045128 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl" Dec 04 13:30:04 crc kubenswrapper[4979]: I1204 13:30:04.687720 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b"] Dec 04 13:30:04 crc kubenswrapper[4979]: I1204 13:30:04.699928 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414205-m4m2b"] Dec 04 13:30:06 crc kubenswrapper[4979]: I1204 13:30:06.210427 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0551bbc-e978-423a-8210-1f99102787a7" path="/var/lib/kubelet/pods/d0551bbc-e978-423a-8210-1f99102787a7/volumes" Dec 04 13:30:07 crc kubenswrapper[4979]: I1204 13:30:07.199652 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:30:07 crc kubenswrapper[4979]: E1204 13:30:07.200254 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:30:08 crc kubenswrapper[4979]: I1204 13:30:08.949194 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h4hd7"] Dec 04 13:30:08 crc kubenswrapper[4979]: E1204 13:30:08.950127 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d314ef80-82e1-4a37-9e06-340d1ca94e89" containerName="collect-profiles" Dec 04 13:30:08 crc kubenswrapper[4979]: I1204 13:30:08.950140 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d314ef80-82e1-4a37-9e06-340d1ca94e89" containerName="collect-profiles" Dec 04 13:30:08 crc kubenswrapper[4979]: I1204 13:30:08.950491 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d314ef80-82e1-4a37-9e06-340d1ca94e89" containerName="collect-profiles" Dec 04 13:30:08 crc kubenswrapper[4979]: I1204 13:30:08.954937 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:08 crc kubenswrapper[4979]: I1204 13:30:08.959726 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h4hd7"] Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.128173 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj7mm\" (UniqueName: \"kubernetes.io/projected/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-kube-api-access-hj7mm\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.128641 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-utilities\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.128802 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-catalog-content\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.230445 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj7mm\" (UniqueName: \"kubernetes.io/projected/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-kube-api-access-hj7mm\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.230505 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-utilities\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.230621 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-catalog-content\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.231339 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-catalog-content\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.231577 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-utilities\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.254892 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj7mm\" (UniqueName: \"kubernetes.io/projected/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-kube-api-access-hj7mm\") pod \"community-operators-h4hd7\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.291545 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:09 crc kubenswrapper[4979]: I1204 13:30:09.802739 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h4hd7"] Dec 04 13:30:10 crc kubenswrapper[4979]: I1204 13:30:10.104109 4979 generic.go:334] "Generic (PLEG): container finished" podID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerID="ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5" exitCode=0 Dec 04 13:30:10 crc kubenswrapper[4979]: I1204 13:30:10.104215 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerDied","Data":"ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5"} Dec 04 13:30:10 crc kubenswrapper[4979]: I1204 13:30:10.104415 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerStarted","Data":"463437f71e5ea865c7a3b441dbc799526f590701722cb51b69020cb6b91ae70d"} Dec 04 13:30:11 crc kubenswrapper[4979]: I1204 13:30:11.120499 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerStarted","Data":"b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685"} Dec 04 13:30:11 crc kubenswrapper[4979]: I1204 13:30:11.812964 4979 scope.go:117] "RemoveContainer" containerID="98a9830f316748de1b45673ea3faad97c57602e0787223babafd4e8d47228034" Dec 04 13:30:12 crc kubenswrapper[4979]: I1204 13:30:12.259431 4979 scope.go:117] "RemoveContainer" containerID="22e6dbd3a9145842ea2090c72f861e332e0096c39dc9f53687a22e2321681a28" Dec 04 13:30:12 crc kubenswrapper[4979]: I1204 13:30:12.291054 4979 scope.go:117] "RemoveContainer" containerID="de9062458b15276a25c92d4889ff2351a826d3665770cbe19b7918dac8f98a8e" Dec 04 13:30:13 crc kubenswrapper[4979]: I1204 13:30:13.149208 4979 generic.go:334] "Generic (PLEG): container finished" podID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerID="b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685" exitCode=0 Dec 04 13:30:13 crc kubenswrapper[4979]: I1204 13:30:13.150180 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerDied","Data":"b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685"} Dec 04 13:30:14 crc kubenswrapper[4979]: I1204 13:30:14.161216 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerStarted","Data":"606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce"} Dec 04 13:30:14 crc kubenswrapper[4979]: I1204 13:30:14.196005 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h4hd7" podStartSLOduration=2.6961205230000003 podStartE2EDuration="6.19597959s" podCreationTimestamp="2025-12-04 13:30:08 +0000 UTC" firstStartedPulling="2025-12-04 13:30:10.106072162 +0000 UTC m=+6434.380367976" lastFinishedPulling="2025-12-04 13:30:13.605931239 +0000 UTC m=+6437.880227043" observedRunningTime="2025-12-04 13:30:14.185575221 +0000 UTC m=+6438.459871025" watchObservedRunningTime="2025-12-04 13:30:14.19597959 +0000 UTC m=+6438.470275404" Dec 04 13:30:18 crc kubenswrapper[4979]: I1204 13:30:18.199061 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:30:18 crc kubenswrapper[4979]: E1204 13:30:18.199872 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:30:19 crc kubenswrapper[4979]: I1204 13:30:19.291758 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:19 crc kubenswrapper[4979]: I1204 13:30:19.292126 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:19 crc kubenswrapper[4979]: I1204 13:30:19.338134 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:20 crc kubenswrapper[4979]: I1204 13:30:20.276987 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:20 crc kubenswrapper[4979]: I1204 13:30:20.331189 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h4hd7"] Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.241122 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h4hd7" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="registry-server" containerID="cri-o://606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce" gracePeriod=2 Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.736601 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.836992 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-utilities\") pod \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.837729 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-catalog-content\") pod \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.838047 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj7mm\" (UniqueName: \"kubernetes.io/projected/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-kube-api-access-hj7mm\") pod \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\" (UID: \"fa66506b-2e4b-45c7-aa1b-5e0fed05366e\") " Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.838145 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-utilities" (OuterVolumeSpecName: "utilities") pod "fa66506b-2e4b-45c7-aa1b-5e0fed05366e" (UID: "fa66506b-2e4b-45c7-aa1b-5e0fed05366e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.838908 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.843263 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-kube-api-access-hj7mm" (OuterVolumeSpecName: "kube-api-access-hj7mm") pod "fa66506b-2e4b-45c7-aa1b-5e0fed05366e" (UID: "fa66506b-2e4b-45c7-aa1b-5e0fed05366e"). InnerVolumeSpecName "kube-api-access-hj7mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.889953 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa66506b-2e4b-45c7-aa1b-5e0fed05366e" (UID: "fa66506b-2e4b-45c7-aa1b-5e0fed05366e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.941642 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj7mm\" (UniqueName: \"kubernetes.io/projected/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-kube-api-access-hj7mm\") on node \"crc\" DevicePath \"\"" Dec 04 13:30:22 crc kubenswrapper[4979]: I1204 13:30:22.941690 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa66506b-2e4b-45c7-aa1b-5e0fed05366e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.254682 4979 generic.go:334] "Generic (PLEG): container finished" podID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerID="606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce" exitCode=0 Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.254757 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h4hd7" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.254765 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerDied","Data":"606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce"} Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.256156 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h4hd7" event={"ID":"fa66506b-2e4b-45c7-aa1b-5e0fed05366e","Type":"ContainerDied","Data":"463437f71e5ea865c7a3b441dbc799526f590701722cb51b69020cb6b91ae70d"} Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.256205 4979 scope.go:117] "RemoveContainer" containerID="606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.289705 4979 scope.go:117] "RemoveContainer" containerID="b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.291032 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h4hd7"] Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.303802 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h4hd7"] Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.323925 4979 scope.go:117] "RemoveContainer" containerID="ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.370978 4979 scope.go:117] "RemoveContainer" containerID="606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce" Dec 04 13:30:23 crc kubenswrapper[4979]: E1204 13:30:23.371638 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce\": container with ID starting with 606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce not found: ID does not exist" containerID="606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.371668 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce"} err="failed to get container status \"606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce\": rpc error: code = NotFound desc = could not find container \"606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce\": container with ID starting with 606830d69a499313ff04665727940f3ac5e50cf62afd385093b0949142b1c3ce not found: ID does not exist" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.371687 4979 scope.go:117] "RemoveContainer" containerID="b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685" Dec 04 13:30:23 crc kubenswrapper[4979]: E1204 13:30:23.371988 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685\": container with ID starting with b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685 not found: ID does not exist" containerID="b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.372031 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685"} err="failed to get container status \"b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685\": rpc error: code = NotFound desc = could not find container \"b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685\": container with ID starting with b67ce90628d7c9e4dfc060825bc0b0fc9b148e810cf16a4265bd0d6be0441685 not found: ID does not exist" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.372056 4979 scope.go:117] "RemoveContainer" containerID="ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5" Dec 04 13:30:23 crc kubenswrapper[4979]: E1204 13:30:23.372412 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5\": container with ID starting with ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5 not found: ID does not exist" containerID="ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5" Dec 04 13:30:23 crc kubenswrapper[4979]: I1204 13:30:23.372433 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5"} err="failed to get container status \"ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5\": rpc error: code = NotFound desc = could not find container \"ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5\": container with ID starting with ffbc297eb56f02ac6be9be29582407c2ddd70c9d3168bfb6cd6e3b4c1238c0c5 not found: ID does not exist" Dec 04 13:30:24 crc kubenswrapper[4979]: I1204 13:30:24.213815 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" path="/var/lib/kubelet/pods/fa66506b-2e4b-45c7-aa1b-5e0fed05366e/volumes" Dec 04 13:30:30 crc kubenswrapper[4979]: I1204 13:30:30.198812 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:30:30 crc kubenswrapper[4979]: E1204 13:30:30.199846 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:30:34 crc kubenswrapper[4979]: I1204 13:30:34.043838 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-create-flkrp"] Dec 04 13:30:34 crc kubenswrapper[4979]: I1204 13:30:34.054548 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-create-flkrp"] Dec 04 13:30:34 crc kubenswrapper[4979]: I1204 13:30:34.210776 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac4aba8f-0d93-48bc-b772-2a09e06ac886" path="/var/lib/kubelet/pods/ac4aba8f-0d93-48bc-b772-2a09e06ac886/volumes" Dec 04 13:30:36 crc kubenswrapper[4979]: I1204 13:30:36.031770 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-e8f1-account-create-update-s56mn"] Dec 04 13:30:36 crc kubenswrapper[4979]: I1204 13:30:36.043723 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-e8f1-account-create-update-s56mn"] Dec 04 13:30:36 crc kubenswrapper[4979]: I1204 13:30:36.214184 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43df5618-f033-4aa9-9076-9f9a4dd1363b" path="/var/lib/kubelet/pods/43df5618-f033-4aa9-9076-9f9a4dd1363b/volumes" Dec 04 13:30:41 crc kubenswrapper[4979]: I1204 13:30:41.031599 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-persistence-db-create-stmlb"] Dec 04 13:30:41 crc kubenswrapper[4979]: I1204 13:30:41.043954 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-persistence-db-create-stmlb"] Dec 04 13:30:42 crc kubenswrapper[4979]: I1204 13:30:42.031323 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-b7de-account-create-update-s64mx"] Dec 04 13:30:42 crc kubenswrapper[4979]: I1204 13:30:42.040248 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-b7de-account-create-update-s64mx"] Dec 04 13:30:42 crc kubenswrapper[4979]: I1204 13:30:42.213476 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38a03b33-a912-4006-9247-24dec637cf71" path="/var/lib/kubelet/pods/38a03b33-a912-4006-9247-24dec637cf71/volumes" Dec 04 13:30:42 crc kubenswrapper[4979]: I1204 13:30:42.214394 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5182298-e0b7-4a7e-bb57-eda965d9a041" path="/var/lib/kubelet/pods/c5182298-e0b7-4a7e-bb57-eda965d9a041/volumes" Dec 04 13:30:43 crc kubenswrapper[4979]: I1204 13:30:43.199253 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:30:43 crc kubenswrapper[4979]: E1204 13:30:43.199630 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:30:58 crc kubenswrapper[4979]: I1204 13:30:58.199064 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:30:58 crc kubenswrapper[4979]: E1204 13:30:58.199963 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:31:09 crc kubenswrapper[4979]: I1204 13:31:09.199263 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:31:09 crc kubenswrapper[4979]: E1204 13:31:09.199990 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.123107 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t4c9f"] Dec 04 13:31:12 crc kubenswrapper[4979]: E1204 13:31:12.124097 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="extract-content" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.124115 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="extract-content" Dec 04 13:31:12 crc kubenswrapper[4979]: E1204 13:31:12.124139 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="extract-utilities" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.124147 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="extract-utilities" Dec 04 13:31:12 crc kubenswrapper[4979]: E1204 13:31:12.124169 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="registry-server" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.124177 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="registry-server" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.124429 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa66506b-2e4b-45c7-aa1b-5e0fed05366e" containerName="registry-server" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.127933 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.139515 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t4c9f"] Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.184020 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc9wf\" (UniqueName: \"kubernetes.io/projected/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-kube-api-access-lc9wf\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.184120 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.184145 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-utilities\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.287082 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc9wf\" (UniqueName: \"kubernetes.io/projected/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-kube-api-access-lc9wf\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.287196 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.287222 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-utilities\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.287897 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.288070 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-utilities\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.309073 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc9wf\" (UniqueName: \"kubernetes.io/projected/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-kube-api-access-lc9wf\") pod \"redhat-operators-t4c9f\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.403865 4979 scope.go:117] "RemoveContainer" containerID="5022f8314541adfdca2a69cdf8f131da56377026a2efb1db45231519a5f31d5c" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.433179 4979 scope.go:117] "RemoveContainer" containerID="8060200e439848054f284863dd785c266481d069d02cba0e7562cafe93c5cf90" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.452613 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.495808 4979 scope.go:117] "RemoveContainer" containerID="d5512089d199aeeeb850b0c3f379fc454214bbc3f9b317f3ccee3b1a7cf68bd4" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.621161 4979 scope.go:117] "RemoveContainer" containerID="1a47303949bda25c349ddfd0c163c6692fb6d8b90efa78d1505a076be0f545e8" Dec 04 13:31:12 crc kubenswrapper[4979]: I1204 13:31:12.954222 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t4c9f"] Dec 04 13:31:12 crc kubenswrapper[4979]: W1204 13:31:12.958496 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0742f25f_9a2e_4f1b_9bbb_5a92a8de1537.slice/crio-f5ec9d7a49ea266ce73fdb82f216893e6086a8e3d94123c81454fcac68870681 WatchSource:0}: Error finding container f5ec9d7a49ea266ce73fdb82f216893e6086a8e3d94123c81454fcac68870681: Status 404 returned error can't find the container with id f5ec9d7a49ea266ce73fdb82f216893e6086a8e3d94123c81454fcac68870681 Dec 04 13:31:13 crc kubenswrapper[4979]: I1204 13:31:13.964318 4979 generic.go:334] "Generic (PLEG): container finished" podID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerID="8bb74345e12de791226165eca4a34610aec7db96d1edbdc7f3d39ae13265d8e1" exitCode=0 Dec 04 13:31:13 crc kubenswrapper[4979]: I1204 13:31:13.964662 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerDied","Data":"8bb74345e12de791226165eca4a34610aec7db96d1edbdc7f3d39ae13265d8e1"} Dec 04 13:31:13 crc kubenswrapper[4979]: I1204 13:31:13.964706 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerStarted","Data":"f5ec9d7a49ea266ce73fdb82f216893e6086a8e3d94123c81454fcac68870681"} Dec 04 13:31:14 crc kubenswrapper[4979]: I1204 13:31:14.977237 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerStarted","Data":"daa2a3cf588c9a46ab6ce262d3826f543b13f1c841cb28434663769df38715c1"} Dec 04 13:31:20 crc kubenswrapper[4979]: I1204 13:31:20.032108 4979 generic.go:334] "Generic (PLEG): container finished" podID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerID="daa2a3cf588c9a46ab6ce262d3826f543b13f1c841cb28434663769df38715c1" exitCode=0 Dec 04 13:31:20 crc kubenswrapper[4979]: I1204 13:31:20.032172 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerDied","Data":"daa2a3cf588c9a46ab6ce262d3826f543b13f1c841cb28434663769df38715c1"} Dec 04 13:31:21 crc kubenswrapper[4979]: I1204 13:31:21.044689 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerStarted","Data":"b5345cec2c83e556809e8fb9be2a14b49dc9bb447c8042144ee821b38fa63bb7"} Dec 04 13:31:21 crc kubenswrapper[4979]: I1204 13:31:21.070195 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t4c9f" podStartSLOduration=2.59007032 podStartE2EDuration="9.07017295s" podCreationTimestamp="2025-12-04 13:31:12 +0000 UTC" firstStartedPulling="2025-12-04 13:31:13.96690426 +0000 UTC m=+6498.241200064" lastFinishedPulling="2025-12-04 13:31:20.44700689 +0000 UTC m=+6504.721302694" observedRunningTime="2025-12-04 13:31:21.064221771 +0000 UTC m=+6505.338517585" watchObservedRunningTime="2025-12-04 13:31:21.07017295 +0000 UTC m=+6505.344468754" Dec 04 13:31:22 crc kubenswrapper[4979]: I1204 13:31:22.199192 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:31:22 crc kubenswrapper[4979]: E1204 13:31:22.199895 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:31:22 crc kubenswrapper[4979]: I1204 13:31:22.452972 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:22 crc kubenswrapper[4979]: I1204 13:31:22.453019 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:23 crc kubenswrapper[4979]: I1204 13:31:23.498785 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t4c9f" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="registry-server" probeResult="failure" output=< Dec 04 13:31:23 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 13:31:23 crc kubenswrapper[4979]: > Dec 04 13:31:29 crc kubenswrapper[4979]: I1204 13:31:29.043827 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/octavia-db-sync-pscsr"] Dec 04 13:31:29 crc kubenswrapper[4979]: I1204 13:31:29.060141 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/octavia-db-sync-pscsr"] Dec 04 13:31:30 crc kubenswrapper[4979]: I1204 13:31:30.239700 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cef70eb-2406-49ad-9fe3-83a6786082b2" path="/var/lib/kubelet/pods/7cef70eb-2406-49ad-9fe3-83a6786082b2/volumes" Dec 04 13:31:32 crc kubenswrapper[4979]: I1204 13:31:32.505071 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:32 crc kubenswrapper[4979]: I1204 13:31:32.563117 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:32 crc kubenswrapper[4979]: I1204 13:31:32.745806 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t4c9f"] Dec 04 13:31:34 crc kubenswrapper[4979]: I1204 13:31:34.183402 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t4c9f" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="registry-server" containerID="cri-o://b5345cec2c83e556809e8fb9be2a14b49dc9bb447c8042144ee821b38fa63bb7" gracePeriod=2 Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.197980 4979 generic.go:334] "Generic (PLEG): container finished" podID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerID="b5345cec2c83e556809e8fb9be2a14b49dc9bb447c8042144ee821b38fa63bb7" exitCode=0 Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.198162 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerDied","Data":"b5345cec2c83e556809e8fb9be2a14b49dc9bb447c8042144ee821b38fa63bb7"} Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.396793 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.485182 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc9wf\" (UniqueName: \"kubernetes.io/projected/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-kube-api-access-lc9wf\") pod \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.485663 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-utilities\") pod \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.485832 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content\") pod \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.486534 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-utilities" (OuterVolumeSpecName: "utilities") pod "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" (UID: "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.529918 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-kube-api-access-lc9wf" (OuterVolumeSpecName: "kube-api-access-lc9wf") pod "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" (UID: "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537"). InnerVolumeSpecName "kube-api-access-lc9wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.587762 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" (UID: "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.588413 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content\") pod \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\" (UID: \"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537\") " Dec 04 13:31:35 crc kubenswrapper[4979]: W1204 13:31:35.588627 4979 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537/volumes/kubernetes.io~empty-dir/catalog-content Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.588652 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" (UID: "0742f25f-9a2e-4f1b-9bbb-5a92a8de1537"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.589207 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc9wf\" (UniqueName: \"kubernetes.io/projected/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-kube-api-access-lc9wf\") on node \"crc\" DevicePath \"\"" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.589227 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:31:35 crc kubenswrapper[4979]: I1204 13:31:35.589237 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.206657 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:31:36 crc kubenswrapper[4979]: E1204 13:31:36.206941 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.211175 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t4c9f" Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.211644 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t4c9f" event={"ID":"0742f25f-9a2e-4f1b-9bbb-5a92a8de1537","Type":"ContainerDied","Data":"f5ec9d7a49ea266ce73fdb82f216893e6086a8e3d94123c81454fcac68870681"} Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.211696 4979 scope.go:117] "RemoveContainer" containerID="b5345cec2c83e556809e8fb9be2a14b49dc9bb447c8042144ee821b38fa63bb7" Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.238709 4979 scope.go:117] "RemoveContainer" containerID="daa2a3cf588c9a46ab6ce262d3826f543b13f1c841cb28434663769df38715c1" Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.292231 4979 scope.go:117] "RemoveContainer" containerID="8bb74345e12de791226165eca4a34610aec7db96d1edbdc7f3d39ae13265d8e1" Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.304395 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t4c9f"] Dec 04 13:31:36 crc kubenswrapper[4979]: I1204 13:31:36.314115 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t4c9f"] Dec 04 13:31:38 crc kubenswrapper[4979]: I1204 13:31:38.209957 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" path="/var/lib/kubelet/pods/0742f25f-9a2e-4f1b-9bbb-5a92a8de1537/volumes" Dec 04 13:31:47 crc kubenswrapper[4979]: I1204 13:31:47.199645 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:31:47 crc kubenswrapper[4979]: E1204 13:31:47.200655 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:31:58 crc kubenswrapper[4979]: I1204 13:31:58.199037 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:31:58 crc kubenswrapper[4979]: E1204 13:31:58.199871 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.335841 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k54hn"] Dec 04 13:32:06 crc kubenswrapper[4979]: E1204 13:32:06.336886 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="registry-server" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.336904 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="registry-server" Dec 04 13:32:06 crc kubenswrapper[4979]: E1204 13:32:06.336919 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="extract-utilities" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.336928 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="extract-utilities" Dec 04 13:32:06 crc kubenswrapper[4979]: E1204 13:32:06.336951 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="extract-content" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.336960 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="extract-content" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.337263 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0742f25f-9a2e-4f1b-9bbb-5a92a8de1537" containerName="registry-server" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.339183 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.348807 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k54hn"] Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.397457 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrmhk\" (UniqueName: \"kubernetes.io/projected/e65a06dc-9164-4cd2-9134-f6561e76d388-kube-api-access-nrmhk\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.397885 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-catalog-content\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.397993 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-utilities\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.500103 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-catalog-content\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.500426 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-utilities\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.500498 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrmhk\" (UniqueName: \"kubernetes.io/projected/e65a06dc-9164-4cd2-9134-f6561e76d388-kube-api-access-nrmhk\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.500940 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-utilities\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.500996 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-catalog-content\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.517874 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrmhk\" (UniqueName: \"kubernetes.io/projected/e65a06dc-9164-4cd2-9134-f6561e76d388-kube-api-access-nrmhk\") pod \"redhat-marketplace-k54hn\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:06 crc kubenswrapper[4979]: I1204 13:32:06.677834 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:07 crc kubenswrapper[4979]: I1204 13:32:07.172426 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k54hn"] Dec 04 13:32:07 crc kubenswrapper[4979]: I1204 13:32:07.492146 4979 generic.go:334] "Generic (PLEG): container finished" podID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerID="9938ae7d949f394ef3c22ff23882dfe3a98c322a477fe27b3fc1d3415a1f7e04" exitCode=0 Dec 04 13:32:07 crc kubenswrapper[4979]: I1204 13:32:07.492348 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerDied","Data":"9938ae7d949f394ef3c22ff23882dfe3a98c322a477fe27b3fc1d3415a1f7e04"} Dec 04 13:32:07 crc kubenswrapper[4979]: I1204 13:32:07.492520 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerStarted","Data":"eb258cbf7fd2b580fbd58b1cc5229fdd984743ea93715a2ad151563b64d819b9"} Dec 04 13:32:08 crc kubenswrapper[4979]: I1204 13:32:08.504052 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerStarted","Data":"c64b7078a311e792631f8879168d1d6c30dbb0efce995c2f1a9315c2291c6d92"} Dec 04 13:32:09 crc kubenswrapper[4979]: I1204 13:32:09.515080 4979 generic.go:334] "Generic (PLEG): container finished" podID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerID="c64b7078a311e792631f8879168d1d6c30dbb0efce995c2f1a9315c2291c6d92" exitCode=0 Dec 04 13:32:09 crc kubenswrapper[4979]: I1204 13:32:09.515435 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerDied","Data":"c64b7078a311e792631f8879168d1d6c30dbb0efce995c2f1a9315c2291c6d92"} Dec 04 13:32:10 crc kubenswrapper[4979]: I1204 13:32:10.525630 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerStarted","Data":"f06fb4865639d08036df8584e984a78e5880222d8c7d1a3583de216da6611dc1"} Dec 04 13:32:12 crc kubenswrapper[4979]: I1204 13:32:12.200884 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:32:12 crc kubenswrapper[4979]: E1204 13:32:12.201501 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:32:12 crc kubenswrapper[4979]: I1204 13:32:12.804525 4979 scope.go:117] "RemoveContainer" containerID="5d522e1988c1bca659fd5bd74fff73c7129e131fc9cf8f562dc05d6b8f9c374b" Dec 04 13:32:12 crc kubenswrapper[4979]: I1204 13:32:12.849132 4979 scope.go:117] "RemoveContainer" containerID="7f0cfdbe507a575fd67f39aba66b7a5a2539f999829a35e10c7d82742cc3490a" Dec 04 13:32:16 crc kubenswrapper[4979]: I1204 13:32:16.678113 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:16 crc kubenswrapper[4979]: I1204 13:32:16.678527 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:16 crc kubenswrapper[4979]: I1204 13:32:16.729918 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:16 crc kubenswrapper[4979]: I1204 13:32:16.746612 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k54hn" podStartSLOduration=8.329961889 podStartE2EDuration="10.7465932s" podCreationTimestamp="2025-12-04 13:32:06 +0000 UTC" firstStartedPulling="2025-12-04 13:32:07.494719707 +0000 UTC m=+6551.769015511" lastFinishedPulling="2025-12-04 13:32:09.911351018 +0000 UTC m=+6554.185646822" observedRunningTime="2025-12-04 13:32:10.550712852 +0000 UTC m=+6554.825008676" watchObservedRunningTime="2025-12-04 13:32:16.7465932 +0000 UTC m=+6561.020889014" Dec 04 13:32:17 crc kubenswrapper[4979]: I1204 13:32:17.639254 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:17 crc kubenswrapper[4979]: I1204 13:32:17.685520 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k54hn"] Dec 04 13:32:19 crc kubenswrapper[4979]: I1204 13:32:19.614359 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k54hn" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="registry-server" containerID="cri-o://f06fb4865639d08036df8584e984a78e5880222d8c7d1a3583de216da6611dc1" gracePeriod=2 Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.631574 4979 generic.go:334] "Generic (PLEG): container finished" podID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerID="f06fb4865639d08036df8584e984a78e5880222d8c7d1a3583de216da6611dc1" exitCode=0 Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.631646 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerDied","Data":"f06fb4865639d08036df8584e984a78e5880222d8c7d1a3583de216da6611dc1"} Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.632164 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k54hn" event={"ID":"e65a06dc-9164-4cd2-9134-f6561e76d388","Type":"ContainerDied","Data":"eb258cbf7fd2b580fbd58b1cc5229fdd984743ea93715a2ad151563b64d819b9"} Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.632181 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb258cbf7fd2b580fbd58b1cc5229fdd984743ea93715a2ad151563b64d819b9" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.685043 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.723046 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-utilities\") pod \"e65a06dc-9164-4cd2-9134-f6561e76d388\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.723250 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrmhk\" (UniqueName: \"kubernetes.io/projected/e65a06dc-9164-4cd2-9134-f6561e76d388-kube-api-access-nrmhk\") pod \"e65a06dc-9164-4cd2-9134-f6561e76d388\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.723328 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-catalog-content\") pod \"e65a06dc-9164-4cd2-9134-f6561e76d388\" (UID: \"e65a06dc-9164-4cd2-9134-f6561e76d388\") " Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.724204 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-utilities" (OuterVolumeSpecName: "utilities") pod "e65a06dc-9164-4cd2-9134-f6561e76d388" (UID: "e65a06dc-9164-4cd2-9134-f6561e76d388"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.729651 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e65a06dc-9164-4cd2-9134-f6561e76d388-kube-api-access-nrmhk" (OuterVolumeSpecName: "kube-api-access-nrmhk") pod "e65a06dc-9164-4cd2-9134-f6561e76d388" (UID: "e65a06dc-9164-4cd2-9134-f6561e76d388"). InnerVolumeSpecName "kube-api-access-nrmhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.755352 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e65a06dc-9164-4cd2-9134-f6561e76d388" (UID: "e65a06dc-9164-4cd2-9134-f6561e76d388"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.826224 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.826270 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrmhk\" (UniqueName: \"kubernetes.io/projected/e65a06dc-9164-4cd2-9134-f6561e76d388-kube-api-access-nrmhk\") on node \"crc\" DevicePath \"\"" Dec 04 13:32:20 crc kubenswrapper[4979]: I1204 13:32:20.826283 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e65a06dc-9164-4cd2-9134-f6561e76d388-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:32:21 crc kubenswrapper[4979]: I1204 13:32:21.642090 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k54hn" Dec 04 13:32:21 crc kubenswrapper[4979]: I1204 13:32:21.678740 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k54hn"] Dec 04 13:32:21 crc kubenswrapper[4979]: I1204 13:32:21.688098 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k54hn"] Dec 04 13:32:22 crc kubenswrapper[4979]: I1204 13:32:22.213477 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" path="/var/lib/kubelet/pods/e65a06dc-9164-4cd2-9134-f6561e76d388/volumes" Dec 04 13:32:25 crc kubenswrapper[4979]: I1204 13:32:25.199995 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:32:25 crc kubenswrapper[4979]: E1204 13:32:25.201202 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:32:39 crc kubenswrapper[4979]: I1204 13:32:39.199397 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:32:39 crc kubenswrapper[4979]: E1204 13:32:39.200167 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:32:51 crc kubenswrapper[4979]: I1204 13:32:51.199501 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:32:51 crc kubenswrapper[4979]: E1204 13:32:51.200251 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:33:05 crc kubenswrapper[4979]: I1204 13:33:05.199089 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:33:05 crc kubenswrapper[4979]: E1204 13:33:05.199849 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:33:18 crc kubenswrapper[4979]: I1204 13:33:18.199689 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:33:18 crc kubenswrapper[4979]: E1204 13:33:18.200411 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:33:33 crc kubenswrapper[4979]: I1204 13:33:33.199578 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:33:33 crc kubenswrapper[4979]: E1204 13:33:33.200167 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:33:48 crc kubenswrapper[4979]: I1204 13:33:48.198540 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:33:48 crc kubenswrapper[4979]: E1204 13:33:48.199230 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:34:01 crc kubenswrapper[4979]: I1204 13:34:01.053190 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-2czth"] Dec 04 13:34:01 crc kubenswrapper[4979]: I1204 13:34:01.063610 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-6ad5-account-create-update-7p225"] Dec 04 13:34:01 crc kubenswrapper[4979]: I1204 13:34:01.073571 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-2czth"] Dec 04 13:34:01 crc kubenswrapper[4979]: I1204 13:34:01.084619 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-6ad5-account-create-update-7p225"] Dec 04 13:34:01 crc kubenswrapper[4979]: I1204 13:34:01.199115 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:34:01 crc kubenswrapper[4979]: E1204 13:34:01.199560 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:34:02 crc kubenswrapper[4979]: I1204 13:34:02.210700 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="055c74a7-df67-4434-960e-e0deab4e5827" path="/var/lib/kubelet/pods/055c74a7-df67-4434-960e-e0deab4e5827/volumes" Dec 04 13:34:02 crc kubenswrapper[4979]: I1204 13:34:02.211643 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47eb1d7d-1fe4-4491-8584-b667a830621a" path="/var/lib/kubelet/pods/47eb1d7d-1fe4-4491-8584-b667a830621a/volumes" Dec 04 13:34:13 crc kubenswrapper[4979]: I1204 13:34:13.018564 4979 scope.go:117] "RemoveContainer" containerID="fa84032727b1245143fdb17593f307cbc0738908e870c1446d6cd3e0a52425d4" Dec 04 13:34:13 crc kubenswrapper[4979]: I1204 13:34:13.043978 4979 scope.go:117] "RemoveContainer" containerID="299093ef32438cc3517c379c1ee45df6194fe9f79c468445dcee7a57fb7bf1ed" Dec 04 13:34:14 crc kubenswrapper[4979]: I1204 13:34:14.033182 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-nvnkz"] Dec 04 13:34:14 crc kubenswrapper[4979]: I1204 13:34:14.043793 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-nvnkz"] Dec 04 13:34:14 crc kubenswrapper[4979]: I1204 13:34:14.212787 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b321d456-d89e-45f2-aa4a-807af0bc5028" path="/var/lib/kubelet/pods/b321d456-d89e-45f2-aa4a-807af0bc5028/volumes" Dec 04 13:34:16 crc kubenswrapper[4979]: I1204 13:34:16.209922 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:34:16 crc kubenswrapper[4979]: E1204 13:34:16.210532 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:34:31 crc kubenswrapper[4979]: I1204 13:34:31.199729 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:34:31 crc kubenswrapper[4979]: I1204 13:34:31.867614 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"6399ff0fa3a2f238b7deb293370a489b954ff58e21bc7f3afd73b489003f776c"} Dec 04 13:35:13 crc kubenswrapper[4979]: I1204 13:35:13.189367 4979 scope.go:117] "RemoveContainer" containerID="2b2fc19afe3e94939e6780da1a95b4aee8d4904f5e72c45ca9885074347fa37d" Dec 04 13:36:27 crc kubenswrapper[4979]: I1204 13:36:27.042724 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-2e13-account-create-update-z9rxp"] Dec 04 13:36:27 crc kubenswrapper[4979]: I1204 13:36:27.051536 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-rddkh"] Dec 04 13:36:27 crc kubenswrapper[4979]: I1204 13:36:27.063653 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-2e13-account-create-update-z9rxp"] Dec 04 13:36:27 crc kubenswrapper[4979]: I1204 13:36:27.071618 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-rddkh"] Dec 04 13:36:28 crc kubenswrapper[4979]: I1204 13:36:28.225278 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="548a7b59-3465-4157-b1f7-ac66e3e7aa6a" path="/var/lib/kubelet/pods/548a7b59-3465-4157-b1f7-ac66e3e7aa6a/volumes" Dec 04 13:36:28 crc kubenswrapper[4979]: I1204 13:36:28.226563 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a" path="/var/lib/kubelet/pods/99a4d56b-c6b2-4574-b5b6-e8cbe7938c4a/volumes" Dec 04 13:36:39 crc kubenswrapper[4979]: I1204 13:36:39.031590 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-cvwdx"] Dec 04 13:36:39 crc kubenswrapper[4979]: I1204 13:36:39.042500 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-cvwdx"] Dec 04 13:36:40 crc kubenswrapper[4979]: I1204 13:36:40.210386 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44910178-dbfd-4f8d-8e28-4357a1927185" path="/var/lib/kubelet/pods/44910178-dbfd-4f8d-8e28-4357a1927185/volumes" Dec 04 13:36:58 crc kubenswrapper[4979]: I1204 13:36:58.040910 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:36:58 crc kubenswrapper[4979]: I1204 13:36:58.042004 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:37:01 crc kubenswrapper[4979]: I1204 13:37:01.047773 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-ce1f-account-create-update-lkvfb"] Dec 04 13:37:01 crc kubenswrapper[4979]: I1204 13:37:01.061789 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-cwpc4"] Dec 04 13:37:01 crc kubenswrapper[4979]: I1204 13:37:01.073063 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-ce1f-account-create-update-lkvfb"] Dec 04 13:37:01 crc kubenswrapper[4979]: I1204 13:37:01.083149 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-cwpc4"] Dec 04 13:37:02 crc kubenswrapper[4979]: I1204 13:37:02.210351 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="140ae347-58c1-49a2-9ff1-9eb734203a83" path="/var/lib/kubelet/pods/140ae347-58c1-49a2-9ff1-9eb734203a83/volumes" Dec 04 13:37:02 crc kubenswrapper[4979]: I1204 13:37:02.211589 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="429bb3c3-eb68-4baf-9ab1-380fcfa50351" path="/var/lib/kubelet/pods/429bb3c3-eb68-4baf-9ab1-380fcfa50351/volumes" Dec 04 13:37:11 crc kubenswrapper[4979]: I1204 13:37:11.028925 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-ztz4s"] Dec 04 13:37:11 crc kubenswrapper[4979]: I1204 13:37:11.040529 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-ztz4s"] Dec 04 13:37:12 crc kubenswrapper[4979]: I1204 13:37:12.212688 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a57214b-fa12-4ac5-bc3a-773b80a03a19" path="/var/lib/kubelet/pods/0a57214b-fa12-4ac5-bc3a-773b80a03a19/volumes" Dec 04 13:37:13 crc kubenswrapper[4979]: I1204 13:37:13.285876 4979 scope.go:117] "RemoveContainer" containerID="99d7f20537bca087ccb36855fb78791136e48ba785d0af1ea8d1a921f402a28d" Dec 04 13:37:13 crc kubenswrapper[4979]: I1204 13:37:13.315029 4979 scope.go:117] "RemoveContainer" containerID="0509d520f85761adb2c0cd1029774dc64f537e2bc2c34f0d2081c76c6591af52" Dec 04 13:37:13 crc kubenswrapper[4979]: I1204 13:37:13.370561 4979 scope.go:117] "RemoveContainer" containerID="cb325e1503e264ba4e72615bfe4b6bf6c3027229f31d700e663bac7437836313" Dec 04 13:37:13 crc kubenswrapper[4979]: I1204 13:37:13.421386 4979 scope.go:117] "RemoveContainer" containerID="ced6b82964cee69bcb2fe7620c8f3061e5b13172a108dba3925934d8a6a23bd1" Dec 04 13:37:13 crc kubenswrapper[4979]: I1204 13:37:13.471148 4979 scope.go:117] "RemoveContainer" containerID="7d7fe9d7db5308742dd3e78daab1f0a1d44d39e2c67b0e6c62d6a1d5ed8dc40d" Dec 04 13:37:13 crc kubenswrapper[4979]: I1204 13:37:13.534992 4979 scope.go:117] "RemoveContainer" containerID="1dbaac49665d193b86af97e3f435c947e6c70b729ece2244804081bea071060c" Dec 04 13:37:28 crc kubenswrapper[4979]: I1204 13:37:28.040883 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:37:28 crc kubenswrapper[4979]: I1204 13:37:28.041474 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.040588 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.041233 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.041292 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.042140 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6399ff0fa3a2f238b7deb293370a489b954ff58e21bc7f3afd73b489003f776c"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.042212 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://6399ff0fa3a2f238b7deb293370a489b954ff58e21bc7f3afd73b489003f776c" gracePeriod=600 Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.965814 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="6399ff0fa3a2f238b7deb293370a489b954ff58e21bc7f3afd73b489003f776c" exitCode=0 Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.965861 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"6399ff0fa3a2f238b7deb293370a489b954ff58e21bc7f3afd73b489003f776c"} Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.966667 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d"} Dec 04 13:37:58 crc kubenswrapper[4979]: I1204 13:37:58.966693 4979 scope.go:117] "RemoveContainer" containerID="ee810d817a8abdd23073684d7bce40ab149835a57d39ff0500534142c7a987a7" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.586552 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8fkgm"] Dec 04 13:38:10 crc kubenswrapper[4979]: E1204 13:38:10.587718 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="extract-utilities" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.587736 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="extract-utilities" Dec 04 13:38:10 crc kubenswrapper[4979]: E1204 13:38:10.587753 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="registry-server" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.587761 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="registry-server" Dec 04 13:38:10 crc kubenswrapper[4979]: E1204 13:38:10.587782 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="extract-content" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.587790 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="extract-content" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.588057 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e65a06dc-9164-4cd2-9134-f6561e76d388" containerName="registry-server" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.592122 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.606663 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8fkgm"] Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.647931 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-catalog-content\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.648114 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtgpl\" (UniqueName: \"kubernetes.io/projected/dc4dd3c2-f037-461e-a689-da825fbf7e5f-kube-api-access-rtgpl\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.648317 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-utilities\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.751439 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-catalog-content\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.751525 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtgpl\" (UniqueName: \"kubernetes.io/projected/dc4dd3c2-f037-461e-a689-da825fbf7e5f-kube-api-access-rtgpl\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.751585 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-utilities\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.752057 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-catalog-content\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.752217 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-utilities\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.771817 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtgpl\" (UniqueName: \"kubernetes.io/projected/dc4dd3c2-f037-461e-a689-da825fbf7e5f-kube-api-access-rtgpl\") pod \"certified-operators-8fkgm\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:10 crc kubenswrapper[4979]: I1204 13:38:10.924245 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:11 crc kubenswrapper[4979]: I1204 13:38:11.469163 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8fkgm"] Dec 04 13:38:12 crc kubenswrapper[4979]: I1204 13:38:12.107035 4979 generic.go:334] "Generic (PLEG): container finished" podID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerID="fc8a9c9ea11d91645d079158785df8eccc563231f93404ae14276e32f52db636" exitCode=0 Dec 04 13:38:12 crc kubenswrapper[4979]: I1204 13:38:12.107389 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerDied","Data":"fc8a9c9ea11d91645d079158785df8eccc563231f93404ae14276e32f52db636"} Dec 04 13:38:12 crc kubenswrapper[4979]: I1204 13:38:12.107423 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerStarted","Data":"3e5492036996326b9d345c5eebc5ca162f9491b6e14a5f4f6d750a570334f7d5"} Dec 04 13:38:12 crc kubenswrapper[4979]: I1204 13:38:12.109827 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:38:13 crc kubenswrapper[4979]: I1204 13:38:13.119465 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerStarted","Data":"ea956fdd2b4a629e4f2fc87220208320821b0a2f75935893f8e396205a0558ad"} Dec 04 13:38:13 crc kubenswrapper[4979]: I1204 13:38:13.735371 4979 scope.go:117] "RemoveContainer" containerID="f06fb4865639d08036df8584e984a78e5880222d8c7d1a3583de216da6611dc1" Dec 04 13:38:13 crc kubenswrapper[4979]: I1204 13:38:13.818097 4979 scope.go:117] "RemoveContainer" containerID="9938ae7d949f394ef3c22ff23882dfe3a98c322a477fe27b3fc1d3415a1f7e04" Dec 04 13:38:13 crc kubenswrapper[4979]: I1204 13:38:13.839499 4979 scope.go:117] "RemoveContainer" containerID="c64b7078a311e792631f8879168d1d6c30dbb0efce995c2f1a9315c2291c6d92" Dec 04 13:38:14 crc kubenswrapper[4979]: I1204 13:38:14.135788 4979 generic.go:334] "Generic (PLEG): container finished" podID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerID="ea956fdd2b4a629e4f2fc87220208320821b0a2f75935893f8e396205a0558ad" exitCode=0 Dec 04 13:38:14 crc kubenswrapper[4979]: I1204 13:38:14.136086 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerDied","Data":"ea956fdd2b4a629e4f2fc87220208320821b0a2f75935893f8e396205a0558ad"} Dec 04 13:38:16 crc kubenswrapper[4979]: I1204 13:38:16.156467 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerStarted","Data":"3c95ef97ae4a660e6688020228670b13fd86ae2de57393a32d9ea420e67ae30b"} Dec 04 13:38:16 crc kubenswrapper[4979]: I1204 13:38:16.184266 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8fkgm" podStartSLOduration=3.071644732 podStartE2EDuration="6.184246385s" podCreationTimestamp="2025-12-04 13:38:10 +0000 UTC" firstStartedPulling="2025-12-04 13:38:12.109617586 +0000 UTC m=+6916.383913390" lastFinishedPulling="2025-12-04 13:38:15.222219239 +0000 UTC m=+6919.496515043" observedRunningTime="2025-12-04 13:38:16.175930442 +0000 UTC m=+6920.450226266" watchObservedRunningTime="2025-12-04 13:38:16.184246385 +0000 UTC m=+6920.458542189" Dec 04 13:38:20 crc kubenswrapper[4979]: I1204 13:38:20.925010 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:20 crc kubenswrapper[4979]: I1204 13:38:20.925590 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:20 crc kubenswrapper[4979]: I1204 13:38:20.996183 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:21 crc kubenswrapper[4979]: I1204 13:38:21.293364 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:21 crc kubenswrapper[4979]: I1204 13:38:21.349652 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8fkgm"] Dec 04 13:38:23 crc kubenswrapper[4979]: I1204 13:38:23.228737 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8fkgm" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="registry-server" containerID="cri-o://3c95ef97ae4a660e6688020228670b13fd86ae2de57393a32d9ea420e67ae30b" gracePeriod=2 Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.240124 4979 generic.go:334] "Generic (PLEG): container finished" podID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerID="3c95ef97ae4a660e6688020228670b13fd86ae2de57393a32d9ea420e67ae30b" exitCode=0 Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.240202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerDied","Data":"3c95ef97ae4a660e6688020228670b13fd86ae2de57393a32d9ea420e67ae30b"} Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.241003 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8fkgm" event={"ID":"dc4dd3c2-f037-461e-a689-da825fbf7e5f","Type":"ContainerDied","Data":"3e5492036996326b9d345c5eebc5ca162f9491b6e14a5f4f6d750a570334f7d5"} Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.241051 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e5492036996326b9d345c5eebc5ca162f9491b6e14a5f4f6d750a570334f7d5" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.266977 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.444700 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-catalog-content\") pod \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.444962 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-utilities\") pod \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.445080 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtgpl\" (UniqueName: \"kubernetes.io/projected/dc4dd3c2-f037-461e-a689-da825fbf7e5f-kube-api-access-rtgpl\") pod \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\" (UID: \"dc4dd3c2-f037-461e-a689-da825fbf7e5f\") " Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.445960 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-utilities" (OuterVolumeSpecName: "utilities") pod "dc4dd3c2-f037-461e-a689-da825fbf7e5f" (UID: "dc4dd3c2-f037-461e-a689-da825fbf7e5f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.453130 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc4dd3c2-f037-461e-a689-da825fbf7e5f-kube-api-access-rtgpl" (OuterVolumeSpecName: "kube-api-access-rtgpl") pod "dc4dd3c2-f037-461e-a689-da825fbf7e5f" (UID: "dc4dd3c2-f037-461e-a689-da825fbf7e5f"). InnerVolumeSpecName "kube-api-access-rtgpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.487448 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dc4dd3c2-f037-461e-a689-da825fbf7e5f" (UID: "dc4dd3c2-f037-461e-a689-da825fbf7e5f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.548098 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.548134 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtgpl\" (UniqueName: \"kubernetes.io/projected/dc4dd3c2-f037-461e-a689-da825fbf7e5f-kube-api-access-rtgpl\") on node \"crc\" DevicePath \"\"" Dec 04 13:38:24 crc kubenswrapper[4979]: I1204 13:38:24.548146 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dc4dd3c2-f037-461e-a689-da825fbf7e5f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:38:25 crc kubenswrapper[4979]: I1204 13:38:25.251846 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8fkgm" Dec 04 13:38:25 crc kubenswrapper[4979]: I1204 13:38:25.303540 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8fkgm"] Dec 04 13:38:25 crc kubenswrapper[4979]: I1204 13:38:25.313539 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8fkgm"] Dec 04 13:38:26 crc kubenswrapper[4979]: I1204 13:38:26.235791 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" path="/var/lib/kubelet/pods/dc4dd3c2-f037-461e-a689-da825fbf7e5f/volumes" Dec 04 13:39:58 crc kubenswrapper[4979]: I1204 13:39:58.040950 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:39:58 crc kubenswrapper[4979]: I1204 13:39:58.041490 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:40:11 crc kubenswrapper[4979]: I1204 13:40:11.273518 4979 generic.go:334] "Generic (PLEG): container finished" podID="717bd66f-9118-4b81-927c-27f66aa6483a" containerID="b8ee396e67fd425c0385bcfc89c8fd5029a4e7a17d3b9056fb54518252de39df" exitCode=0 Dec 04 13:40:11 crc kubenswrapper[4979]: I1204 13:40:11.273670 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" event={"ID":"717bd66f-9118-4b81-927c-27f66aa6483a","Type":"ContainerDied","Data":"b8ee396e67fd425c0385bcfc89c8fd5029a4e7a17d3b9056fb54518252de39df"} Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.752176 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.821854 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ceph\") pod \"717bd66f-9118-4b81-927c-27f66aa6483a\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.821988 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dbnv\" (UniqueName: \"kubernetes.io/projected/717bd66f-9118-4b81-927c-27f66aa6483a-kube-api-access-2dbnv\") pod \"717bd66f-9118-4b81-927c-27f66aa6483a\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.822067 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-tripleo-cleanup-combined-ca-bundle\") pod \"717bd66f-9118-4b81-927c-27f66aa6483a\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.822146 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ssh-key\") pod \"717bd66f-9118-4b81-927c-27f66aa6483a\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.822320 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-inventory\") pod \"717bd66f-9118-4b81-927c-27f66aa6483a\" (UID: \"717bd66f-9118-4b81-927c-27f66aa6483a\") " Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.827263 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/717bd66f-9118-4b81-927c-27f66aa6483a-kube-api-access-2dbnv" (OuterVolumeSpecName: "kube-api-access-2dbnv") pod "717bd66f-9118-4b81-927c-27f66aa6483a" (UID: "717bd66f-9118-4b81-927c-27f66aa6483a"). InnerVolumeSpecName "kube-api-access-2dbnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.828547 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ceph" (OuterVolumeSpecName: "ceph") pod "717bd66f-9118-4b81-927c-27f66aa6483a" (UID: "717bd66f-9118-4b81-927c-27f66aa6483a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.845236 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-tripleo-cleanup-combined-ca-bundle" (OuterVolumeSpecName: "tripleo-cleanup-combined-ca-bundle") pod "717bd66f-9118-4b81-927c-27f66aa6483a" (UID: "717bd66f-9118-4b81-927c-27f66aa6483a"). InnerVolumeSpecName "tripleo-cleanup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.852183 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-inventory" (OuterVolumeSpecName: "inventory") pod "717bd66f-9118-4b81-927c-27f66aa6483a" (UID: "717bd66f-9118-4b81-927c-27f66aa6483a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.856585 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "717bd66f-9118-4b81-927c-27f66aa6483a" (UID: "717bd66f-9118-4b81-927c-27f66aa6483a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.924861 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dbnv\" (UniqueName: \"kubernetes.io/projected/717bd66f-9118-4b81-927c-27f66aa6483a-kube-api-access-2dbnv\") on node \"crc\" DevicePath \"\"" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.924917 4979 reconciler_common.go:293] "Volume detached for volume \"tripleo-cleanup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-tripleo-cleanup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.924931 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.924940 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:40:12 crc kubenswrapper[4979]: I1204 13:40:12.924951 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/717bd66f-9118-4b81-927c-27f66aa6483a-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:40:13 crc kubenswrapper[4979]: I1204 13:40:13.301127 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" event={"ID":"717bd66f-9118-4b81-927c-27f66aa6483a","Type":"ContainerDied","Data":"280645de966a386ce45e465834df8c048b34f4d8c169dd7aea6d63d08708cf72"} Dec 04 13:40:13 crc kubenswrapper[4979]: I1204 13:40:13.301186 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="280645de966a386ce45e465834df8c048b34f4d8c169dd7aea6d63d08708cf72" Dec 04 13:40:13 crc kubenswrapper[4979]: I1204 13:40:13.301256 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.726655 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-4f44q"] Dec 04 13:40:16 crc kubenswrapper[4979]: E1204 13:40:16.728061 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="extract-utilities" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.728084 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="extract-utilities" Dec 04 13:40:16 crc kubenswrapper[4979]: E1204 13:40:16.728110 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="717bd66f-9118-4b81-927c-27f66aa6483a" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.728126 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="717bd66f-9118-4b81-927c-27f66aa6483a" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 04 13:40:16 crc kubenswrapper[4979]: E1204 13:40:16.728173 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="registry-server" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.728185 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="registry-server" Dec 04 13:40:16 crc kubenswrapper[4979]: E1204 13:40:16.728224 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="extract-content" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.728236 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="extract-content" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.728660 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc4dd3c2-f037-461e-a689-da825fbf7e5f" containerName="registry-server" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.728708 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="717bd66f-9118-4b81-927c-27f66aa6483a" containerName="tripleo-cleanup-tripleo-cleanup-openstack-cell1" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.729977 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.732977 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.733032 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.733078 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.733113 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.739169 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-4f44q"] Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.809015 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.809097 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fmsc\" (UniqueName: \"kubernetes.io/projected/902a1159-bb56-419e-ac6f-9cb05257d54f-kube-api-access-7fmsc\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.809227 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ceph\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.809270 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.809516 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-inventory\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.912770 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.913400 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fmsc\" (UniqueName: \"kubernetes.io/projected/902a1159-bb56-419e-ac6f-9cb05257d54f-kube-api-access-7fmsc\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.913544 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ceph\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.913584 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.913723 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-inventory\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.919573 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-inventory\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.919606 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ceph\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.920888 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ssh-key\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.921386 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-bootstrap-combined-ca-bundle\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:16 crc kubenswrapper[4979]: I1204 13:40:16.942577 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fmsc\" (UniqueName: \"kubernetes.io/projected/902a1159-bb56-419e-ac6f-9cb05257d54f-kube-api-access-7fmsc\") pod \"bootstrap-openstack-openstack-cell1-4f44q\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:17 crc kubenswrapper[4979]: I1204 13:40:17.056907 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:40:17 crc kubenswrapper[4979]: I1204 13:40:17.629784 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-openstack-openstack-cell1-4f44q"] Dec 04 13:40:18 crc kubenswrapper[4979]: I1204 13:40:18.356938 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" event={"ID":"902a1159-bb56-419e-ac6f-9cb05257d54f","Type":"ContainerStarted","Data":"65bf5e6f8cda581433f5227d56e3ef452195ddf5825e6ec612a1fa24d0d85bab"} Dec 04 13:40:18 crc kubenswrapper[4979]: I1204 13:40:18.358004 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" event={"ID":"902a1159-bb56-419e-ac6f-9cb05257d54f","Type":"ContainerStarted","Data":"e0bb77d19073cf3dff94ed9edc5db3d06e328094a3e8adbd59c7aeb01823fea6"} Dec 04 13:40:18 crc kubenswrapper[4979]: I1204 13:40:18.385191 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" podStartSLOduration=1.944927722 podStartE2EDuration="2.385173468s" podCreationTimestamp="2025-12-04 13:40:16 +0000 UTC" firstStartedPulling="2025-12-04 13:40:17.636330888 +0000 UTC m=+7041.910626682" lastFinishedPulling="2025-12-04 13:40:18.076576624 +0000 UTC m=+7042.350872428" observedRunningTime="2025-12-04 13:40:18.377081371 +0000 UTC m=+7042.651377175" watchObservedRunningTime="2025-12-04 13:40:18.385173468 +0000 UTC m=+7042.659469272" Dec 04 13:40:28 crc kubenswrapper[4979]: I1204 13:40:28.040445 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:40:28 crc kubenswrapper[4979]: I1204 13:40:28.041086 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.040894 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.041343 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.041381 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.042118 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.042160 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" gracePeriod=600 Dec 04 13:40:58 crc kubenswrapper[4979]: E1204 13:40:58.164511 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.806218 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" exitCode=0 Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.806483 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d"} Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.806723 4979 scope.go:117] "RemoveContainer" containerID="6399ff0fa3a2f238b7deb293370a489b954ff58e21bc7f3afd73b489003f776c" Dec 04 13:40:58 crc kubenswrapper[4979]: I1204 13:40:58.807891 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:40:58 crc kubenswrapper[4979]: E1204 13:40:58.808575 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:41:11 crc kubenswrapper[4979]: I1204 13:41:11.200469 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:41:11 crc kubenswrapper[4979]: E1204 13:41:11.202024 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.160895 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-j5tfk"] Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.163615 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.177126 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j5tfk"] Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.320073 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-utilities\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.320589 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbj7k\" (UniqueName: \"kubernetes.io/projected/8834e175-fb15-469c-9775-51d76d293dd2-kube-api-access-cbj7k\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.321072 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-catalog-content\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.423437 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbj7k\" (UniqueName: \"kubernetes.io/projected/8834e175-fb15-469c-9775-51d76d293dd2-kube-api-access-cbj7k\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.423595 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-catalog-content\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.423676 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-utilities\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.424222 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-utilities\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.424289 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-catalog-content\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.445002 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbj7k\" (UniqueName: \"kubernetes.io/projected/8834e175-fb15-469c-9775-51d76d293dd2-kube-api-access-cbj7k\") pod \"community-operators-j5tfk\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:15 crc kubenswrapper[4979]: I1204 13:41:15.491387 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:16 crc kubenswrapper[4979]: I1204 13:41:16.007806 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-j5tfk"] Dec 04 13:41:16 crc kubenswrapper[4979]: I1204 13:41:16.999265 4979 generic.go:334] "Generic (PLEG): container finished" podID="8834e175-fb15-469c-9775-51d76d293dd2" containerID="b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23" exitCode=0 Dec 04 13:41:16 crc kubenswrapper[4979]: I1204 13:41:16.999413 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerDied","Data":"b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23"} Dec 04 13:41:17 crc kubenswrapper[4979]: I1204 13:41:16.999902 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerStarted","Data":"4b95a1fcceb11b0fcc05baad44b26f6790a67108e644cf4c7ddbe2a3c91e1948"} Dec 04 13:41:18 crc kubenswrapper[4979]: I1204 13:41:18.011557 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerStarted","Data":"32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a"} Dec 04 13:41:19 crc kubenswrapper[4979]: I1204 13:41:19.024406 4979 generic.go:334] "Generic (PLEG): container finished" podID="8834e175-fb15-469c-9775-51d76d293dd2" containerID="32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a" exitCode=0 Dec 04 13:41:19 crc kubenswrapper[4979]: I1204 13:41:19.024559 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerDied","Data":"32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a"} Dec 04 13:41:21 crc kubenswrapper[4979]: I1204 13:41:21.050875 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerStarted","Data":"c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d"} Dec 04 13:41:21 crc kubenswrapper[4979]: I1204 13:41:21.087599 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-j5tfk" podStartSLOduration=3.286148774 podStartE2EDuration="6.087570026s" podCreationTimestamp="2025-12-04 13:41:15 +0000 UTC" firstStartedPulling="2025-12-04 13:41:17.001693756 +0000 UTC m=+7101.275989560" lastFinishedPulling="2025-12-04 13:41:19.803115008 +0000 UTC m=+7104.077410812" observedRunningTime="2025-12-04 13:41:21.076873838 +0000 UTC m=+7105.351169652" watchObservedRunningTime="2025-12-04 13:41:21.087570026 +0000 UTC m=+7105.361865840" Dec 04 13:41:24 crc kubenswrapper[4979]: I1204 13:41:24.200570 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:41:24 crc kubenswrapper[4979]: E1204 13:41:24.202420 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:41:25 crc kubenswrapper[4979]: I1204 13:41:25.491702 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:25 crc kubenswrapper[4979]: I1204 13:41:25.492106 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:25 crc kubenswrapper[4979]: I1204 13:41:25.539404 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:26 crc kubenswrapper[4979]: I1204 13:41:26.146286 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:26 crc kubenswrapper[4979]: I1204 13:41:26.195151 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j5tfk"] Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.112539 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-j5tfk" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="registry-server" containerID="cri-o://c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d" gracePeriod=2 Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.600972 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.717103 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-utilities\") pod \"8834e175-fb15-469c-9775-51d76d293dd2\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.717219 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbj7k\" (UniqueName: \"kubernetes.io/projected/8834e175-fb15-469c-9775-51d76d293dd2-kube-api-access-cbj7k\") pod \"8834e175-fb15-469c-9775-51d76d293dd2\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.717246 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-catalog-content\") pod \"8834e175-fb15-469c-9775-51d76d293dd2\" (UID: \"8834e175-fb15-469c-9775-51d76d293dd2\") " Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.718378 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-utilities" (OuterVolumeSpecName: "utilities") pod "8834e175-fb15-469c-9775-51d76d293dd2" (UID: "8834e175-fb15-469c-9775-51d76d293dd2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.723566 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8834e175-fb15-469c-9775-51d76d293dd2-kube-api-access-cbj7k" (OuterVolumeSpecName: "kube-api-access-cbj7k") pod "8834e175-fb15-469c-9775-51d76d293dd2" (UID: "8834e175-fb15-469c-9775-51d76d293dd2"). InnerVolumeSpecName "kube-api-access-cbj7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.771043 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8834e175-fb15-469c-9775-51d76d293dd2" (UID: "8834e175-fb15-469c-9775-51d76d293dd2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.819840 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.819875 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbj7k\" (UniqueName: \"kubernetes.io/projected/8834e175-fb15-469c-9775-51d76d293dd2-kube-api-access-cbj7k\") on node \"crc\" DevicePath \"\"" Dec 04 13:41:28 crc kubenswrapper[4979]: I1204 13:41:28.819886 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8834e175-fb15-469c-9775-51d76d293dd2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.124659 4979 generic.go:334] "Generic (PLEG): container finished" podID="8834e175-fb15-469c-9775-51d76d293dd2" containerID="c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d" exitCode=0 Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.124721 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-j5tfk" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.124751 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerDied","Data":"c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d"} Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.125038 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-j5tfk" event={"ID":"8834e175-fb15-469c-9775-51d76d293dd2","Type":"ContainerDied","Data":"4b95a1fcceb11b0fcc05baad44b26f6790a67108e644cf4c7ddbe2a3c91e1948"} Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.125062 4979 scope.go:117] "RemoveContainer" containerID="c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.156665 4979 scope.go:117] "RemoveContainer" containerID="32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.157692 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-j5tfk"] Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.177082 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-j5tfk"] Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.179287 4979 scope.go:117] "RemoveContainer" containerID="b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.236327 4979 scope.go:117] "RemoveContainer" containerID="c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d" Dec 04 13:41:29 crc kubenswrapper[4979]: E1204 13:41:29.236779 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d\": container with ID starting with c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d not found: ID does not exist" containerID="c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.236810 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d"} err="failed to get container status \"c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d\": rpc error: code = NotFound desc = could not find container \"c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d\": container with ID starting with c93c385aff2d665890d4c0454eecd9f9dc2dff79371a807b4582fb60ab8aa06d not found: ID does not exist" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.236834 4979 scope.go:117] "RemoveContainer" containerID="32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a" Dec 04 13:41:29 crc kubenswrapper[4979]: E1204 13:41:29.237548 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a\": container with ID starting with 32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a not found: ID does not exist" containerID="32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.237575 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a"} err="failed to get container status \"32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a\": rpc error: code = NotFound desc = could not find container \"32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a\": container with ID starting with 32117d9f35eda353383b4daca746865779144c91102ad615244b573e043be52a not found: ID does not exist" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.237588 4979 scope.go:117] "RemoveContainer" containerID="b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23" Dec 04 13:41:29 crc kubenswrapper[4979]: E1204 13:41:29.237847 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23\": container with ID starting with b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23 not found: ID does not exist" containerID="b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23" Dec 04 13:41:29 crc kubenswrapper[4979]: I1204 13:41:29.237879 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23"} err="failed to get container status \"b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23\": rpc error: code = NotFound desc = could not find container \"b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23\": container with ID starting with b660d93191bea5c25398dcc5126bd210d3ef7f86f4ac7e2d0bf8fb0b6e871d23 not found: ID does not exist" Dec 04 13:41:30 crc kubenswrapper[4979]: I1204 13:41:30.215078 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8834e175-fb15-469c-9775-51d76d293dd2" path="/var/lib/kubelet/pods/8834e175-fb15-469c-9775-51d76d293dd2/volumes" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.890619 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rt927"] Dec 04 13:41:31 crc kubenswrapper[4979]: E1204 13:41:31.891532 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="registry-server" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.891551 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="registry-server" Dec 04 13:41:31 crc kubenswrapper[4979]: E1204 13:41:31.891565 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="extract-content" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.891573 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="extract-content" Dec 04 13:41:31 crc kubenswrapper[4979]: E1204 13:41:31.891593 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="extract-utilities" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.891601 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="extract-utilities" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.891877 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8834e175-fb15-469c-9775-51d76d293dd2" containerName="registry-server" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.893888 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.903715 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rt927"] Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.992725 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-catalog-content\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.992834 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-utilities\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:31 crc kubenswrapper[4979]: I1204 13:41:31.992888 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pj4pl\" (UniqueName: \"kubernetes.io/projected/ea080944-2423-4166-bc26-9d1e5b0d4403-kube-api-access-pj4pl\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.095416 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-catalog-content\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.095751 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-utilities\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.095788 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pj4pl\" (UniqueName: \"kubernetes.io/projected/ea080944-2423-4166-bc26-9d1e5b0d4403-kube-api-access-pj4pl\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.096139 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-utilities\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.096139 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-catalog-content\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.116350 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pj4pl\" (UniqueName: \"kubernetes.io/projected/ea080944-2423-4166-bc26-9d1e5b0d4403-kube-api-access-pj4pl\") pod \"redhat-operators-rt927\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.232243 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:32 crc kubenswrapper[4979]: I1204 13:41:32.722748 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rt927"] Dec 04 13:41:33 crc kubenswrapper[4979]: I1204 13:41:33.168194 4979 generic.go:334] "Generic (PLEG): container finished" podID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerID="57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74" exitCode=0 Dec 04 13:41:33 crc kubenswrapper[4979]: I1204 13:41:33.168390 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerDied","Data":"57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74"} Dec 04 13:41:33 crc kubenswrapper[4979]: I1204 13:41:33.168530 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerStarted","Data":"590bb9372b31dcccb052889e9e4126f1775423d3b214009f38c3406d45797101"} Dec 04 13:41:34 crc kubenswrapper[4979]: I1204 13:41:34.180202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerStarted","Data":"b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13"} Dec 04 13:41:38 crc kubenswrapper[4979]: I1204 13:41:38.199004 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:41:38 crc kubenswrapper[4979]: E1204 13:41:38.200802 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:41:38 crc kubenswrapper[4979]: I1204 13:41:38.217534 4979 generic.go:334] "Generic (PLEG): container finished" podID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerID="b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13" exitCode=0 Dec 04 13:41:38 crc kubenswrapper[4979]: I1204 13:41:38.217585 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerDied","Data":"b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13"} Dec 04 13:41:39 crc kubenswrapper[4979]: I1204 13:41:39.228986 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerStarted","Data":"b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53"} Dec 04 13:41:39 crc kubenswrapper[4979]: I1204 13:41:39.257362 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rt927" podStartSLOduration=2.7164067960000002 podStartE2EDuration="8.257342876s" podCreationTimestamp="2025-12-04 13:41:31 +0000 UTC" firstStartedPulling="2025-12-04 13:41:33.17067801 +0000 UTC m=+7117.444973814" lastFinishedPulling="2025-12-04 13:41:38.71161409 +0000 UTC m=+7122.985909894" observedRunningTime="2025-12-04 13:41:39.247617765 +0000 UTC m=+7123.521913569" watchObservedRunningTime="2025-12-04 13:41:39.257342876 +0000 UTC m=+7123.531638690" Dec 04 13:41:42 crc kubenswrapper[4979]: I1204 13:41:42.232679 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:42 crc kubenswrapper[4979]: I1204 13:41:42.233272 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:43 crc kubenswrapper[4979]: I1204 13:41:43.279412 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-rt927" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="registry-server" probeResult="failure" output=< Dec 04 13:41:43 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 13:41:43 crc kubenswrapper[4979]: > Dec 04 13:41:52 crc kubenswrapper[4979]: I1204 13:41:52.280537 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:52 crc kubenswrapper[4979]: I1204 13:41:52.352694 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:52 crc kubenswrapper[4979]: I1204 13:41:52.519505 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rt927"] Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.200526 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:41:53 crc kubenswrapper[4979]: E1204 13:41:53.201132 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.357591 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rt927" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="registry-server" containerID="cri-o://b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53" gracePeriod=2 Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.918320 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.968708 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-catalog-content\") pod \"ea080944-2423-4166-bc26-9d1e5b0d4403\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.968924 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-utilities\") pod \"ea080944-2423-4166-bc26-9d1e5b0d4403\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.969000 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj4pl\" (UniqueName: \"kubernetes.io/projected/ea080944-2423-4166-bc26-9d1e5b0d4403-kube-api-access-pj4pl\") pod \"ea080944-2423-4166-bc26-9d1e5b0d4403\" (UID: \"ea080944-2423-4166-bc26-9d1e5b0d4403\") " Dec 04 13:41:53 crc kubenswrapper[4979]: I1204 13:41:53.970459 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-utilities" (OuterVolumeSpecName: "utilities") pod "ea080944-2423-4166-bc26-9d1e5b0d4403" (UID: "ea080944-2423-4166-bc26-9d1e5b0d4403"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.008757 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea080944-2423-4166-bc26-9d1e5b0d4403-kube-api-access-pj4pl" (OuterVolumeSpecName: "kube-api-access-pj4pl") pod "ea080944-2423-4166-bc26-9d1e5b0d4403" (UID: "ea080944-2423-4166-bc26-9d1e5b0d4403"). InnerVolumeSpecName "kube-api-access-pj4pl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.071872 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj4pl\" (UniqueName: \"kubernetes.io/projected/ea080944-2423-4166-bc26-9d1e5b0d4403-kube-api-access-pj4pl\") on node \"crc\" DevicePath \"\"" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.071912 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.099888 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ea080944-2423-4166-bc26-9d1e5b0d4403" (UID: "ea080944-2423-4166-bc26-9d1e5b0d4403"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.174536 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ea080944-2423-4166-bc26-9d1e5b0d4403-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.368143 4979 generic.go:334] "Generic (PLEG): container finished" podID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerID="b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53" exitCode=0 Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.368197 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerDied","Data":"b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53"} Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.368230 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rt927" event={"ID":"ea080944-2423-4166-bc26-9d1e5b0d4403","Type":"ContainerDied","Data":"590bb9372b31dcccb052889e9e4126f1775423d3b214009f38c3406d45797101"} Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.368279 4979 scope.go:117] "RemoveContainer" containerID="b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.368290 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rt927" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.391833 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rt927"] Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.392153 4979 scope.go:117] "RemoveContainer" containerID="b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.404366 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rt927"] Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.412594 4979 scope.go:117] "RemoveContainer" containerID="57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.468362 4979 scope.go:117] "RemoveContainer" containerID="b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53" Dec 04 13:41:54 crc kubenswrapper[4979]: E1204 13:41:54.468891 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53\": container with ID starting with b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53 not found: ID does not exist" containerID="b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.468959 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53"} err="failed to get container status \"b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53\": rpc error: code = NotFound desc = could not find container \"b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53\": container with ID starting with b5460c29f8fc45e1ec7673db1e75dbdcbb3c37ba9439c9f1b97319acbe34db53 not found: ID does not exist" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.468984 4979 scope.go:117] "RemoveContainer" containerID="b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13" Dec 04 13:41:54 crc kubenswrapper[4979]: E1204 13:41:54.469580 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13\": container with ID starting with b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13 not found: ID does not exist" containerID="b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.469623 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13"} err="failed to get container status \"b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13\": rpc error: code = NotFound desc = could not find container \"b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13\": container with ID starting with b7c285bb04d3cd4a3f0a063e1c9e9a2e9e3a27ddb021450bab4ed29552e3cf13 not found: ID does not exist" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.469642 4979 scope.go:117] "RemoveContainer" containerID="57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74" Dec 04 13:41:54 crc kubenswrapper[4979]: E1204 13:41:54.470043 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74\": container with ID starting with 57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74 not found: ID does not exist" containerID="57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74" Dec 04 13:41:54 crc kubenswrapper[4979]: I1204 13:41:54.470087 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74"} err="failed to get container status \"57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74\": rpc error: code = NotFound desc = could not find container \"57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74\": container with ID starting with 57abdbb9fd1a42e1a1701493a38c30f7e90a845d5605361f415fed64c9e32d74 not found: ID does not exist" Dec 04 13:41:56 crc kubenswrapper[4979]: I1204 13:41:56.215064 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" path="/var/lib/kubelet/pods/ea080944-2423-4166-bc26-9d1e5b0d4403/volumes" Dec 04 13:42:06 crc kubenswrapper[4979]: I1204 13:42:06.221871 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:42:06 crc kubenswrapper[4979]: E1204 13:42:06.223224 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:42:19 crc kubenswrapper[4979]: I1204 13:42:19.199923 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:42:19 crc kubenswrapper[4979]: E1204 13:42:19.201017 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:42:31 crc kubenswrapper[4979]: I1204 13:42:31.199417 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:42:31 crc kubenswrapper[4979]: E1204 13:42:31.200181 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:42:42 crc kubenswrapper[4979]: I1204 13:42:42.200485 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:42:42 crc kubenswrapper[4979]: E1204 13:42:42.201914 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:42:56 crc kubenswrapper[4979]: I1204 13:42:56.214555 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:42:56 crc kubenswrapper[4979]: E1204 13:42:56.215965 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:43:08 crc kubenswrapper[4979]: I1204 13:43:08.200214 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:43:08 crc kubenswrapper[4979]: E1204 13:43:08.202417 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:43:23 crc kubenswrapper[4979]: I1204 13:43:23.199558 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:43:23 crc kubenswrapper[4979]: E1204 13:43:23.201050 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:43:27 crc kubenswrapper[4979]: I1204 13:43:27.335447 4979 generic.go:334] "Generic (PLEG): container finished" podID="902a1159-bb56-419e-ac6f-9cb05257d54f" containerID="65bf5e6f8cda581433f5227d56e3ef452195ddf5825e6ec612a1fa24d0d85bab" exitCode=0 Dec 04 13:43:27 crc kubenswrapper[4979]: I1204 13:43:27.335542 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" event={"ID":"902a1159-bb56-419e-ac6f-9cb05257d54f","Type":"ContainerDied","Data":"65bf5e6f8cda581433f5227d56e3ef452195ddf5825e6ec612a1fa24d0d85bab"} Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.782579 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.862266 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ceph\") pod \"902a1159-bb56-419e-ac6f-9cb05257d54f\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.862397 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ssh-key\") pod \"902a1159-bb56-419e-ac6f-9cb05257d54f\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.862562 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-inventory\") pod \"902a1159-bb56-419e-ac6f-9cb05257d54f\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.862622 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fmsc\" (UniqueName: \"kubernetes.io/projected/902a1159-bb56-419e-ac6f-9cb05257d54f-kube-api-access-7fmsc\") pod \"902a1159-bb56-419e-ac6f-9cb05257d54f\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.862745 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-bootstrap-combined-ca-bundle\") pod \"902a1159-bb56-419e-ac6f-9cb05257d54f\" (UID: \"902a1159-bb56-419e-ac6f-9cb05257d54f\") " Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.868154 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "902a1159-bb56-419e-ac6f-9cb05257d54f" (UID: "902a1159-bb56-419e-ac6f-9cb05257d54f"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.869055 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902a1159-bb56-419e-ac6f-9cb05257d54f-kube-api-access-7fmsc" (OuterVolumeSpecName: "kube-api-access-7fmsc") pod "902a1159-bb56-419e-ac6f-9cb05257d54f" (UID: "902a1159-bb56-419e-ac6f-9cb05257d54f"). InnerVolumeSpecName "kube-api-access-7fmsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.871345 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ceph" (OuterVolumeSpecName: "ceph") pod "902a1159-bb56-419e-ac6f-9cb05257d54f" (UID: "902a1159-bb56-419e-ac6f-9cb05257d54f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.895490 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-inventory" (OuterVolumeSpecName: "inventory") pod "902a1159-bb56-419e-ac6f-9cb05257d54f" (UID: "902a1159-bb56-419e-ac6f-9cb05257d54f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.899168 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "902a1159-bb56-419e-ac6f-9cb05257d54f" (UID: "902a1159-bb56-419e-ac6f-9cb05257d54f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.966135 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fmsc\" (UniqueName: \"kubernetes.io/projected/902a1159-bb56-419e-ac6f-9cb05257d54f-kube-api-access-7fmsc\") on node \"crc\" DevicePath \"\"" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.966167 4979 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.966177 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.966186 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:43:28 crc kubenswrapper[4979]: I1204 13:43:28.966194 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/902a1159-bb56-419e-ac6f-9cb05257d54f-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.359863 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" event={"ID":"902a1159-bb56-419e-ac6f-9cb05257d54f","Type":"ContainerDied","Data":"e0bb77d19073cf3dff94ed9edc5db3d06e328094a3e8adbd59c7aeb01823fea6"} Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.360157 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e0bb77d19073cf3dff94ed9edc5db3d06e328094a3e8adbd59c7aeb01823fea6" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.359924 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-openstack-openstack-cell1-4f44q" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.464909 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-6dj6r"] Dec 04 13:43:29 crc kubenswrapper[4979]: E1204 13:43:29.465483 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="extract-utilities" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.465509 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="extract-utilities" Dec 04 13:43:29 crc kubenswrapper[4979]: E1204 13:43:29.465529 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="extract-content" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.465539 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="extract-content" Dec 04 13:43:29 crc kubenswrapper[4979]: E1204 13:43:29.465557 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="registry-server" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.465565 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="registry-server" Dec 04 13:43:29 crc kubenswrapper[4979]: E1204 13:43:29.465586 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902a1159-bb56-419e-ac6f-9cb05257d54f" containerName="bootstrap-openstack-openstack-cell1" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.465595 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="902a1159-bb56-419e-ac6f-9cb05257d54f" containerName="bootstrap-openstack-openstack-cell1" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.465854 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="902a1159-bb56-419e-ac6f-9cb05257d54f" containerName="bootstrap-openstack-openstack-cell1" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.465904 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea080944-2423-4166-bc26-9d1e5b0d4403" containerName="registry-server" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.468687 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.473388 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.473937 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.474193 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.474366 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.500240 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-6dj6r"] Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.578574 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ceph\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.578635 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-inventory\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.578668 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ssh-key\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.578714 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5wj9\" (UniqueName: \"kubernetes.io/projected/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-kube-api-access-s5wj9\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.681208 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ceph\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.681258 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-inventory\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.681292 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ssh-key\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.681364 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5wj9\" (UniqueName: \"kubernetes.io/projected/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-kube-api-access-s5wj9\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.685525 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ssh-key\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.685841 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-inventory\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.685843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ceph\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.696928 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5wj9\" (UniqueName: \"kubernetes.io/projected/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-kube-api-access-s5wj9\") pod \"download-cache-openstack-openstack-cell1-6dj6r\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:29 crc kubenswrapper[4979]: I1204 13:43:29.805005 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:43:30 crc kubenswrapper[4979]: I1204 13:43:30.421039 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-openstack-openstack-cell1-6dj6r"] Dec 04 13:43:30 crc kubenswrapper[4979]: I1204 13:43:30.424050 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:43:31 crc kubenswrapper[4979]: I1204 13:43:31.389652 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" event={"ID":"78c8481a-65fe-4e8a-85fc-d55cfb0dd709","Type":"ContainerStarted","Data":"e21b939c62db1fa688f87a41a9a620e6ac6eedea4b7c4a698898336973bbfaac"} Dec 04 13:43:31 crc kubenswrapper[4979]: I1204 13:43:31.390108 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" event={"ID":"78c8481a-65fe-4e8a-85fc-d55cfb0dd709","Type":"ContainerStarted","Data":"3928e3ebdb402b27c6be414ff6402ad7206ec90635c283b5af5afc09dcbc285f"} Dec 04 13:43:31 crc kubenswrapper[4979]: I1204 13:43:31.416183 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" podStartSLOduration=1.98530027 podStartE2EDuration="2.416161305s" podCreationTimestamp="2025-12-04 13:43:29 +0000 UTC" firstStartedPulling="2025-12-04 13:43:30.423840713 +0000 UTC m=+7234.698136517" lastFinishedPulling="2025-12-04 13:43:30.854701718 +0000 UTC m=+7235.128997552" observedRunningTime="2025-12-04 13:43:31.405819178 +0000 UTC m=+7235.680114982" watchObservedRunningTime="2025-12-04 13:43:31.416161305 +0000 UTC m=+7235.690457119" Dec 04 13:43:35 crc kubenswrapper[4979]: I1204 13:43:35.198889 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:43:35 crc kubenswrapper[4979]: E1204 13:43:35.199687 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:43:46 crc kubenswrapper[4979]: I1204 13:43:46.207719 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:43:46 crc kubenswrapper[4979]: E1204 13:43:46.209157 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:43:58 crc kubenswrapper[4979]: I1204 13:43:58.199914 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:43:58 crc kubenswrapper[4979]: E1204 13:43:58.201541 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:44:13 crc kubenswrapper[4979]: I1204 13:44:13.198971 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:44:13 crc kubenswrapper[4979]: E1204 13:44:13.199676 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:44:14 crc kubenswrapper[4979]: I1204 13:44:14.112855 4979 scope.go:117] "RemoveContainer" containerID="ea956fdd2b4a629e4f2fc87220208320821b0a2f75935893f8e396205a0558ad" Dec 04 13:44:14 crc kubenswrapper[4979]: I1204 13:44:14.140071 4979 scope.go:117] "RemoveContainer" containerID="fc8a9c9ea11d91645d079158785df8eccc563231f93404ae14276e32f52db636" Dec 04 13:44:26 crc kubenswrapper[4979]: I1204 13:44:26.207996 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:44:26 crc kubenswrapper[4979]: E1204 13:44:26.208757 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.696445 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bj8sz"] Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.699470 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.705443 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bj8sz"] Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.855741 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-utilities\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.856220 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxtcj\" (UniqueName: \"kubernetes.io/projected/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-kube-api-access-rxtcj\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.856389 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-catalog-content\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.959340 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-utilities\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.959481 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxtcj\" (UniqueName: \"kubernetes.io/projected/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-kube-api-access-rxtcj\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.959540 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-catalog-content\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.960016 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-utilities\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.960087 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-catalog-content\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:36 crc kubenswrapper[4979]: I1204 13:44:36.984269 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxtcj\" (UniqueName: \"kubernetes.io/projected/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-kube-api-access-rxtcj\") pod \"redhat-marketplace-bj8sz\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:37 crc kubenswrapper[4979]: I1204 13:44:37.026909 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:37 crc kubenswrapper[4979]: I1204 13:44:37.491739 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bj8sz"] Dec 04 13:44:38 crc kubenswrapper[4979]: I1204 13:44:38.196368 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerStarted","Data":"2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f"} Dec 04 13:44:38 crc kubenswrapper[4979]: I1204 13:44:38.196674 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerStarted","Data":"5ef1a1d0038c48b949562013c6884dca215c87a7b91327e18d4cc58093ad59d5"} Dec 04 13:44:39 crc kubenswrapper[4979]: I1204 13:44:39.199416 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:44:39 crc kubenswrapper[4979]: E1204 13:44:39.200161 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:44:39 crc kubenswrapper[4979]: I1204 13:44:39.208804 4979 generic.go:334] "Generic (PLEG): container finished" podID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerID="2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f" exitCode=0 Dec 04 13:44:39 crc kubenswrapper[4979]: I1204 13:44:39.208856 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerDied","Data":"2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f"} Dec 04 13:44:41 crc kubenswrapper[4979]: I1204 13:44:41.234361 4979 generic.go:334] "Generic (PLEG): container finished" podID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerID="84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b" exitCode=0 Dec 04 13:44:41 crc kubenswrapper[4979]: I1204 13:44:41.234504 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerDied","Data":"84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b"} Dec 04 13:44:43 crc kubenswrapper[4979]: I1204 13:44:43.264231 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerStarted","Data":"fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e"} Dec 04 13:44:43 crc kubenswrapper[4979]: I1204 13:44:43.306142 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bj8sz" podStartSLOduration=4.500208043 podStartE2EDuration="7.306120094s" podCreationTimestamp="2025-12-04 13:44:36 +0000 UTC" firstStartedPulling="2025-12-04 13:44:39.212033896 +0000 UTC m=+7303.486329700" lastFinishedPulling="2025-12-04 13:44:42.017945907 +0000 UTC m=+7306.292241751" observedRunningTime="2025-12-04 13:44:43.291903503 +0000 UTC m=+7307.566199327" watchObservedRunningTime="2025-12-04 13:44:43.306120094 +0000 UTC m=+7307.580415918" Dec 04 13:44:47 crc kubenswrapper[4979]: I1204 13:44:47.027889 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:47 crc kubenswrapper[4979]: I1204 13:44:47.028664 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:47 crc kubenswrapper[4979]: I1204 13:44:47.087170 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:47 crc kubenswrapper[4979]: I1204 13:44:47.372625 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:47 crc kubenswrapper[4979]: I1204 13:44:47.435147 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bj8sz"] Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.323918 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bj8sz" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="registry-server" containerID="cri-o://fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e" gracePeriod=2 Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.765401 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.859979 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxtcj\" (UniqueName: \"kubernetes.io/projected/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-kube-api-access-rxtcj\") pod \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.860454 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-catalog-content\") pod \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.860522 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-utilities\") pod \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\" (UID: \"6940e982-daf6-4d9d-bc1b-7d0a24869a9f\") " Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.861596 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-utilities" (OuterVolumeSpecName: "utilities") pod "6940e982-daf6-4d9d-bc1b-7d0a24869a9f" (UID: "6940e982-daf6-4d9d-bc1b-7d0a24869a9f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.864715 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-kube-api-access-rxtcj" (OuterVolumeSpecName: "kube-api-access-rxtcj") pod "6940e982-daf6-4d9d-bc1b-7d0a24869a9f" (UID: "6940e982-daf6-4d9d-bc1b-7d0a24869a9f"). InnerVolumeSpecName "kube-api-access-rxtcj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.880422 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6940e982-daf6-4d9d-bc1b-7d0a24869a9f" (UID: "6940e982-daf6-4d9d-bc1b-7d0a24869a9f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.963262 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.963380 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:44:49 crc kubenswrapper[4979]: I1204 13:44:49.963402 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxtcj\" (UniqueName: \"kubernetes.io/projected/6940e982-daf6-4d9d-bc1b-7d0a24869a9f-kube-api-access-rxtcj\") on node \"crc\" DevicePath \"\"" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.340902 4979 generic.go:334] "Generic (PLEG): container finished" podID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerID="fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e" exitCode=0 Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.340958 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerDied","Data":"fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e"} Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.341008 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bj8sz" event={"ID":"6940e982-daf6-4d9d-bc1b-7d0a24869a9f","Type":"ContainerDied","Data":"5ef1a1d0038c48b949562013c6884dca215c87a7b91327e18d4cc58093ad59d5"} Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.341033 4979 scope.go:117] "RemoveContainer" containerID="fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.341041 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bj8sz" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.377856 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bj8sz"] Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.388277 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bj8sz"] Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.397361 4979 scope.go:117] "RemoveContainer" containerID="84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.432647 4979 scope.go:117] "RemoveContainer" containerID="2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.484478 4979 scope.go:117] "RemoveContainer" containerID="fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e" Dec 04 13:44:50 crc kubenswrapper[4979]: E1204 13:44:50.485136 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e\": container with ID starting with fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e not found: ID does not exist" containerID="fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.485198 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e"} err="failed to get container status \"fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e\": rpc error: code = NotFound desc = could not find container \"fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e\": container with ID starting with fc5a8bef7c9451f7c25c0ed15ca10855e2184c8bec4fce8cf17629e5a82a5f6e not found: ID does not exist" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.485241 4979 scope.go:117] "RemoveContainer" containerID="84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b" Dec 04 13:44:50 crc kubenswrapper[4979]: E1204 13:44:50.485845 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b\": container with ID starting with 84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b not found: ID does not exist" containerID="84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.485891 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b"} err="failed to get container status \"84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b\": rpc error: code = NotFound desc = could not find container \"84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b\": container with ID starting with 84bfebd44052708f00dca2ffa6038eee6a9acaf86ac2125320f468971718775b not found: ID does not exist" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.485923 4979 scope.go:117] "RemoveContainer" containerID="2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f" Dec 04 13:44:50 crc kubenswrapper[4979]: E1204 13:44:50.486745 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f\": container with ID starting with 2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f not found: ID does not exist" containerID="2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f" Dec 04 13:44:50 crc kubenswrapper[4979]: I1204 13:44:50.486783 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f"} err="failed to get container status \"2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f\": rpc error: code = NotFound desc = could not find container \"2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f\": container with ID starting with 2102f7c51d149ff3904be7a4aa40ce579b5ddae1dc6596812f5d2b735e9b4d7f not found: ID does not exist" Dec 04 13:44:52 crc kubenswrapper[4979]: I1204 13:44:52.210837 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" path="/var/lib/kubelet/pods/6940e982-daf6-4d9d-bc1b-7d0a24869a9f/volumes" Dec 04 13:44:54 crc kubenswrapper[4979]: I1204 13:44:54.199219 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:44:54 crc kubenswrapper[4979]: E1204 13:44:54.200105 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.155764 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb"] Dec 04 13:45:00 crc kubenswrapper[4979]: E1204 13:45:00.156928 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="extract-content" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.156944 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="extract-content" Dec 04 13:45:00 crc kubenswrapper[4979]: E1204 13:45:00.156963 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="extract-utilities" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.156972 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="extract-utilities" Dec 04 13:45:00 crc kubenswrapper[4979]: E1204 13:45:00.156996 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="registry-server" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.157007 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="registry-server" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.157382 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6940e982-daf6-4d9d-bc1b-7d0a24869a9f" containerName="registry-server" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.158321 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.161586 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.162511 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.165882 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb"] Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.286812 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd899ee9-8952-40e6-b5af-d62e28fe2670-secret-volume\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.286870 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj9gx\" (UniqueName: \"kubernetes.io/projected/fd899ee9-8952-40e6-b5af-d62e28fe2670-kube-api-access-mj9gx\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.288535 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd899ee9-8952-40e6-b5af-d62e28fe2670-config-volume\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.390497 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd899ee9-8952-40e6-b5af-d62e28fe2670-config-volume\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.390623 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd899ee9-8952-40e6-b5af-d62e28fe2670-secret-volume\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.390664 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj9gx\" (UniqueName: \"kubernetes.io/projected/fd899ee9-8952-40e6-b5af-d62e28fe2670-kube-api-access-mj9gx\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.391896 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd899ee9-8952-40e6-b5af-d62e28fe2670-config-volume\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.403947 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd899ee9-8952-40e6-b5af-d62e28fe2670-secret-volume\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.410734 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj9gx\" (UniqueName: \"kubernetes.io/projected/fd899ee9-8952-40e6-b5af-d62e28fe2670-kube-api-access-mj9gx\") pod \"collect-profiles-29414265-rrqzb\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:00 crc kubenswrapper[4979]: I1204 13:45:00.490338 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:01 crc kubenswrapper[4979]: I1204 13:45:01.002759 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb"] Dec 04 13:45:01 crc kubenswrapper[4979]: I1204 13:45:01.453477 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" event={"ID":"fd899ee9-8952-40e6-b5af-d62e28fe2670","Type":"ContainerStarted","Data":"cb94fb6a5d541da09f48ab189b4bfd79abac67e41b34e33033216ee23b0fa9dd"} Dec 04 13:45:01 crc kubenswrapper[4979]: I1204 13:45:01.454547 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" event={"ID":"fd899ee9-8952-40e6-b5af-d62e28fe2670","Type":"ContainerStarted","Data":"17fd37ef0221a52ef1a79d14e65277717617abdb6fbfa1d47707889ad63e5623"} Dec 04 13:45:01 crc kubenswrapper[4979]: I1204 13:45:01.457141 4979 generic.go:334] "Generic (PLEG): container finished" podID="78c8481a-65fe-4e8a-85fc-d55cfb0dd709" containerID="e21b939c62db1fa688f87a41a9a620e6ac6eedea4b7c4a698898336973bbfaac" exitCode=0 Dec 04 13:45:01 crc kubenswrapper[4979]: I1204 13:45:01.457182 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" event={"ID":"78c8481a-65fe-4e8a-85fc-d55cfb0dd709","Type":"ContainerDied","Data":"e21b939c62db1fa688f87a41a9a620e6ac6eedea4b7c4a698898336973bbfaac"} Dec 04 13:45:01 crc kubenswrapper[4979]: I1204 13:45:01.472519 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" podStartSLOduration=1.472442543 podStartE2EDuration="1.472442543s" podCreationTimestamp="2025-12-04 13:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 13:45:01.468518818 +0000 UTC m=+7325.742814652" watchObservedRunningTime="2025-12-04 13:45:01.472442543 +0000 UTC m=+7325.746738347" Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.465783 4979 generic.go:334] "Generic (PLEG): container finished" podID="fd899ee9-8952-40e6-b5af-d62e28fe2670" containerID="cb94fb6a5d541da09f48ab189b4bfd79abac67e41b34e33033216ee23b0fa9dd" exitCode=0 Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.465821 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" event={"ID":"fd899ee9-8952-40e6-b5af-d62e28fe2670","Type":"ContainerDied","Data":"cb94fb6a5d541da09f48ab189b4bfd79abac67e41b34e33033216ee23b0fa9dd"} Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.930910 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.983296 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ceph\") pod \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.983397 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-inventory\") pod \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.983418 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ssh-key\") pod \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.983500 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5wj9\" (UniqueName: \"kubernetes.io/projected/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-kube-api-access-s5wj9\") pod \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\" (UID: \"78c8481a-65fe-4e8a-85fc-d55cfb0dd709\") " Dec 04 13:45:02 crc kubenswrapper[4979]: I1204 13:45:02.991878 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ceph" (OuterVolumeSpecName: "ceph") pod "78c8481a-65fe-4e8a-85fc-d55cfb0dd709" (UID: "78c8481a-65fe-4e8a-85fc-d55cfb0dd709"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.006900 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-kube-api-access-s5wj9" (OuterVolumeSpecName: "kube-api-access-s5wj9") pod "78c8481a-65fe-4e8a-85fc-d55cfb0dd709" (UID: "78c8481a-65fe-4e8a-85fc-d55cfb0dd709"). InnerVolumeSpecName "kube-api-access-s5wj9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.013651 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "78c8481a-65fe-4e8a-85fc-d55cfb0dd709" (UID: "78c8481a-65fe-4e8a-85fc-d55cfb0dd709"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.022952 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-inventory" (OuterVolumeSpecName: "inventory") pod "78c8481a-65fe-4e8a-85fc-d55cfb0dd709" (UID: "78c8481a-65fe-4e8a-85fc-d55cfb0dd709"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.085771 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.085807 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.085819 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5wj9\" (UniqueName: \"kubernetes.io/projected/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-kube-api-access-s5wj9\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.085835 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/78c8481a-65fe-4e8a-85fc-d55cfb0dd709-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.477686 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.477694 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-openstack-openstack-cell1-6dj6r" event={"ID":"78c8481a-65fe-4e8a-85fc-d55cfb0dd709","Type":"ContainerDied","Data":"3928e3ebdb402b27c6be414ff6402ad7206ec90635c283b5af5afc09dcbc285f"} Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.478056 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3928e3ebdb402b27c6be414ff6402ad7206ec90635c283b5af5afc09dcbc285f" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.617608 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-rfcs6"] Dec 04 13:45:03 crc kubenswrapper[4979]: E1204 13:45:03.618261 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78c8481a-65fe-4e8a-85fc-d55cfb0dd709" containerName="download-cache-openstack-openstack-cell1" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.618291 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="78c8481a-65fe-4e8a-85fc-d55cfb0dd709" containerName="download-cache-openstack-openstack-cell1" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.618641 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="78c8481a-65fe-4e8a-85fc-d55cfb0dd709" containerName="download-cache-openstack-openstack-cell1" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.619694 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.625549 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.627657 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.628569 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.629528 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.630760 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-rfcs6"] Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.700722 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-inventory\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.700806 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ceph\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.701033 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ssh-key\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.701087 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgfm4\" (UniqueName: \"kubernetes.io/projected/e81e9d5c-a38e-4340-aca1-548c5b31f652-kube-api-access-jgfm4\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.792385 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.803217 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ssh-key\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.803280 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgfm4\" (UniqueName: \"kubernetes.io/projected/e81e9d5c-a38e-4340-aca1-548c5b31f652-kube-api-access-jgfm4\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.803409 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-inventory\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.803444 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ceph\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.809960 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-inventory\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.810686 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ceph\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.812345 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ssh-key\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.823064 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgfm4\" (UniqueName: \"kubernetes.io/projected/e81e9d5c-a38e-4340-aca1-548c5b31f652-kube-api-access-jgfm4\") pod \"configure-network-openstack-openstack-cell1-rfcs6\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.905139 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd899ee9-8952-40e6-b5af-d62e28fe2670-config-volume\") pod \"fd899ee9-8952-40e6-b5af-d62e28fe2670\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.905458 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj9gx\" (UniqueName: \"kubernetes.io/projected/fd899ee9-8952-40e6-b5af-d62e28fe2670-kube-api-access-mj9gx\") pod \"fd899ee9-8952-40e6-b5af-d62e28fe2670\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.905521 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd899ee9-8952-40e6-b5af-d62e28fe2670-secret-volume\") pod \"fd899ee9-8952-40e6-b5af-d62e28fe2670\" (UID: \"fd899ee9-8952-40e6-b5af-d62e28fe2670\") " Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.906394 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd899ee9-8952-40e6-b5af-d62e28fe2670-config-volume" (OuterVolumeSpecName: "config-volume") pod "fd899ee9-8952-40e6-b5af-d62e28fe2670" (UID: "fd899ee9-8952-40e6-b5af-d62e28fe2670"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.910389 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd899ee9-8952-40e6-b5af-d62e28fe2670-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fd899ee9-8952-40e6-b5af-d62e28fe2670" (UID: "fd899ee9-8952-40e6-b5af-d62e28fe2670"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.940886 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd899ee9-8952-40e6-b5af-d62e28fe2670-kube-api-access-mj9gx" (OuterVolumeSpecName: "kube-api-access-mj9gx") pod "fd899ee9-8952-40e6-b5af-d62e28fe2670" (UID: "fd899ee9-8952-40e6-b5af-d62e28fe2670"). InnerVolumeSpecName "kube-api-access-mj9gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:45:03 crc kubenswrapper[4979]: I1204 13:45:03.951060 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.007838 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj9gx\" (UniqueName: \"kubernetes.io/projected/fd899ee9-8952-40e6-b5af-d62e28fe2670-kube-api-access-mj9gx\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.007872 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fd899ee9-8952-40e6-b5af-d62e28fe2670-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.007882 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fd899ee9-8952-40e6-b5af-d62e28fe2670-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.481394 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-openstack-openstack-cell1-rfcs6"] Dec 04 13:45:04 crc kubenswrapper[4979]: W1204 13:45:04.484330 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode81e9d5c_a38e_4340_aca1_548c5b31f652.slice/crio-31cff21e727b1340bc3712ad5883f6e33c8e526342f352beeb1e8d40ed788c25 WatchSource:0}: Error finding container 31cff21e727b1340bc3712ad5883f6e33c8e526342f352beeb1e8d40ed788c25: Status 404 returned error can't find the container with id 31cff21e727b1340bc3712ad5883f6e33c8e526342f352beeb1e8d40ed788c25 Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.488101 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" event={"ID":"fd899ee9-8952-40e6-b5af-d62e28fe2670","Type":"ContainerDied","Data":"17fd37ef0221a52ef1a79d14e65277717617abdb6fbfa1d47707889ad63e5623"} Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.488138 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17fd37ef0221a52ef1a79d14e65277717617abdb6fbfa1d47707889ad63e5623" Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.488191 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb" Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.551244 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm"] Dec 04 13:45:04 crc kubenswrapper[4979]: I1204 13:45:04.561664 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414220-n2mmm"] Dec 04 13:45:05 crc kubenswrapper[4979]: I1204 13:45:05.502820 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" event={"ID":"e81e9d5c-a38e-4340-aca1-548c5b31f652","Type":"ContainerStarted","Data":"31cff21e727b1340bc3712ad5883f6e33c8e526342f352beeb1e8d40ed788c25"} Dec 04 13:45:06 crc kubenswrapper[4979]: I1204 13:45:06.230924 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154bfa25-b814-486a-808f-c61bcba05297" path="/var/lib/kubelet/pods/154bfa25-b814-486a-808f-c61bcba05297/volumes" Dec 04 13:45:07 crc kubenswrapper[4979]: I1204 13:45:07.521979 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" event={"ID":"e81e9d5c-a38e-4340-aca1-548c5b31f652","Type":"ContainerStarted","Data":"8c758d43d10147f55c0abd44e5e4b78e030ea6dfa219717588bdde8d9a54bcee"} Dec 04 13:45:07 crc kubenswrapper[4979]: I1204 13:45:07.548681 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" podStartSLOduration=2.522143941 podStartE2EDuration="4.548659659s" podCreationTimestamp="2025-12-04 13:45:03 +0000 UTC" firstStartedPulling="2025-12-04 13:45:04.487232436 +0000 UTC m=+7328.761528240" lastFinishedPulling="2025-12-04 13:45:06.513748154 +0000 UTC m=+7330.788043958" observedRunningTime="2025-12-04 13:45:07.539948255 +0000 UTC m=+7331.814244059" watchObservedRunningTime="2025-12-04 13:45:07.548659659 +0000 UTC m=+7331.822955463" Dec 04 13:45:08 crc kubenswrapper[4979]: I1204 13:45:08.206698 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:45:08 crc kubenswrapper[4979]: E1204 13:45:08.207342 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:45:14 crc kubenswrapper[4979]: I1204 13:45:14.219374 4979 scope.go:117] "RemoveContainer" containerID="3c95ef97ae4a660e6688020228670b13fd86ae2de57393a32d9ea420e67ae30b" Dec 04 13:45:14 crc kubenswrapper[4979]: I1204 13:45:14.261549 4979 scope.go:117] "RemoveContainer" containerID="6865f4cd9a4cd0065a5531a535a8e964e4b2383bc7cc2a9ddff8b6fc73785ecc" Dec 04 13:45:21 crc kubenswrapper[4979]: I1204 13:45:21.200092 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:45:21 crc kubenswrapper[4979]: E1204 13:45:21.201019 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:45:33 crc kubenswrapper[4979]: I1204 13:45:33.199436 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:45:33 crc kubenswrapper[4979]: E1204 13:45:33.200509 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:45:48 crc kubenswrapper[4979]: I1204 13:45:48.199616 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:45:48 crc kubenswrapper[4979]: E1204 13:45:48.200462 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:46:01 crc kubenswrapper[4979]: I1204 13:46:01.199925 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:46:02 crc kubenswrapper[4979]: I1204 13:46:02.071730 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"4d83d72a3f1cfee91bd625d5ef34b5bb09b87f3b468acc46be5ce89bc808d8ac"} Dec 04 13:46:30 crc kubenswrapper[4979]: I1204 13:46:30.344671 4979 generic.go:334] "Generic (PLEG): container finished" podID="e81e9d5c-a38e-4340-aca1-548c5b31f652" containerID="8c758d43d10147f55c0abd44e5e4b78e030ea6dfa219717588bdde8d9a54bcee" exitCode=0 Dec 04 13:46:30 crc kubenswrapper[4979]: I1204 13:46:30.344748 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" event={"ID":"e81e9d5c-a38e-4340-aca1-548c5b31f652","Type":"ContainerDied","Data":"8c758d43d10147f55c0abd44e5e4b78e030ea6dfa219717588bdde8d9a54bcee"} Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.813444 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.887265 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-inventory\") pod \"e81e9d5c-a38e-4340-aca1-548c5b31f652\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.887348 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgfm4\" (UniqueName: \"kubernetes.io/projected/e81e9d5c-a38e-4340-aca1-548c5b31f652-kube-api-access-jgfm4\") pod \"e81e9d5c-a38e-4340-aca1-548c5b31f652\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.887424 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ceph\") pod \"e81e9d5c-a38e-4340-aca1-548c5b31f652\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.887579 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ssh-key\") pod \"e81e9d5c-a38e-4340-aca1-548c5b31f652\" (UID: \"e81e9d5c-a38e-4340-aca1-548c5b31f652\") " Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.893534 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81e9d5c-a38e-4340-aca1-548c5b31f652-kube-api-access-jgfm4" (OuterVolumeSpecName: "kube-api-access-jgfm4") pod "e81e9d5c-a38e-4340-aca1-548c5b31f652" (UID: "e81e9d5c-a38e-4340-aca1-548c5b31f652"). InnerVolumeSpecName "kube-api-access-jgfm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.893608 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ceph" (OuterVolumeSpecName: "ceph") pod "e81e9d5c-a38e-4340-aca1-548c5b31f652" (UID: "e81e9d5c-a38e-4340-aca1-548c5b31f652"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.916916 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-inventory" (OuterVolumeSpecName: "inventory") pod "e81e9d5c-a38e-4340-aca1-548c5b31f652" (UID: "e81e9d5c-a38e-4340-aca1-548c5b31f652"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.923595 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e81e9d5c-a38e-4340-aca1-548c5b31f652" (UID: "e81e9d5c-a38e-4340-aca1-548c5b31f652"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.990236 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.990276 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgfm4\" (UniqueName: \"kubernetes.io/projected/e81e9d5c-a38e-4340-aca1-548c5b31f652-kube-api-access-jgfm4\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.990293 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:31 crc kubenswrapper[4979]: I1204 13:46:31.990327 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e81e9d5c-a38e-4340-aca1-548c5b31f652-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.366491 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" event={"ID":"e81e9d5c-a38e-4340-aca1-548c5b31f652","Type":"ContainerDied","Data":"31cff21e727b1340bc3712ad5883f6e33c8e526342f352beeb1e8d40ed788c25"} Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.366797 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31cff21e727b1340bc3712ad5883f6e33c8e526342f352beeb1e8d40ed788c25" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.366601 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-openstack-openstack-cell1-rfcs6" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.462591 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-nq7vm"] Dec 04 13:46:32 crc kubenswrapper[4979]: E1204 13:46:32.463846 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81e9d5c-a38e-4340-aca1-548c5b31f652" containerName="configure-network-openstack-openstack-cell1" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.463885 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81e9d5c-a38e-4340-aca1-548c5b31f652" containerName="configure-network-openstack-openstack-cell1" Dec 04 13:46:32 crc kubenswrapper[4979]: E1204 13:46:32.463936 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd899ee9-8952-40e6-b5af-d62e28fe2670" containerName="collect-profiles" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.463950 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd899ee9-8952-40e6-b5af-d62e28fe2670" containerName="collect-profiles" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.464734 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81e9d5c-a38e-4340-aca1-548c5b31f652" containerName="configure-network-openstack-openstack-cell1" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.464776 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd899ee9-8952-40e6-b5af-d62e28fe2670" containerName="collect-profiles" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.467718 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.471373 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.474574 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.474701 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.474293 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.498720 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-nq7vm"] Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.603326 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ceph\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.603444 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs4qh\" (UniqueName: \"kubernetes.io/projected/8647a121-d252-4072-aded-c1166beefd89-kube-api-access-fs4qh\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.603709 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-inventory\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.603993 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ssh-key\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.705591 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-inventory\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.705708 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ssh-key\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.705842 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ceph\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.705867 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs4qh\" (UniqueName: \"kubernetes.io/projected/8647a121-d252-4072-aded-c1166beefd89-kube-api-access-fs4qh\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.712502 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-inventory\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.712552 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ssh-key\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.713280 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ceph\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.735077 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs4qh\" (UniqueName: \"kubernetes.io/projected/8647a121-d252-4072-aded-c1166beefd89-kube-api-access-fs4qh\") pod \"validate-network-openstack-openstack-cell1-nq7vm\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:32 crc kubenswrapper[4979]: I1204 13:46:32.793714 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:33 crc kubenswrapper[4979]: I1204 13:46:33.353984 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-openstack-openstack-cell1-nq7vm"] Dec 04 13:46:33 crc kubenswrapper[4979]: I1204 13:46:33.380433 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" event={"ID":"8647a121-d252-4072-aded-c1166beefd89","Type":"ContainerStarted","Data":"c1826a691ee10c947f25d41225832bbbec7fe80bf9a84d983d16952aa5702d4d"} Dec 04 13:46:34 crc kubenswrapper[4979]: I1204 13:46:34.396897 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" event={"ID":"8647a121-d252-4072-aded-c1166beefd89","Type":"ContainerStarted","Data":"cdfe079bd0a31326ac62328fc925450f67c275cff90e06c4a28b65f946c6c14a"} Dec 04 13:46:34 crc kubenswrapper[4979]: I1204 13:46:34.421989 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" podStartSLOduration=1.957964678 podStartE2EDuration="2.421968403s" podCreationTimestamp="2025-12-04 13:46:32 +0000 UTC" firstStartedPulling="2025-12-04 13:46:33.364976765 +0000 UTC m=+7417.639272569" lastFinishedPulling="2025-12-04 13:46:33.82898049 +0000 UTC m=+7418.103276294" observedRunningTime="2025-12-04 13:46:34.415229572 +0000 UTC m=+7418.689525376" watchObservedRunningTime="2025-12-04 13:46:34.421968403 +0000 UTC m=+7418.696264217" Dec 04 13:46:39 crc kubenswrapper[4979]: I1204 13:46:39.468815 4979 generic.go:334] "Generic (PLEG): container finished" podID="8647a121-d252-4072-aded-c1166beefd89" containerID="cdfe079bd0a31326ac62328fc925450f67c275cff90e06c4a28b65f946c6c14a" exitCode=0 Dec 04 13:46:39 crc kubenswrapper[4979]: I1204 13:46:39.468930 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" event={"ID":"8647a121-d252-4072-aded-c1166beefd89","Type":"ContainerDied","Data":"cdfe079bd0a31326ac62328fc925450f67c275cff90e06c4a28b65f946c6c14a"} Dec 04 13:46:40 crc kubenswrapper[4979]: I1204 13:46:40.953830 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.100885 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ceph\") pod \"8647a121-d252-4072-aded-c1166beefd89\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.101156 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fs4qh\" (UniqueName: \"kubernetes.io/projected/8647a121-d252-4072-aded-c1166beefd89-kube-api-access-fs4qh\") pod \"8647a121-d252-4072-aded-c1166beefd89\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.101184 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-inventory\") pod \"8647a121-d252-4072-aded-c1166beefd89\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.101272 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ssh-key\") pod \"8647a121-d252-4072-aded-c1166beefd89\" (UID: \"8647a121-d252-4072-aded-c1166beefd89\") " Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.107120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8647a121-d252-4072-aded-c1166beefd89-kube-api-access-fs4qh" (OuterVolumeSpecName: "kube-api-access-fs4qh") pod "8647a121-d252-4072-aded-c1166beefd89" (UID: "8647a121-d252-4072-aded-c1166beefd89"). InnerVolumeSpecName "kube-api-access-fs4qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.107982 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ceph" (OuterVolumeSpecName: "ceph") pod "8647a121-d252-4072-aded-c1166beefd89" (UID: "8647a121-d252-4072-aded-c1166beefd89"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.136473 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-inventory" (OuterVolumeSpecName: "inventory") pod "8647a121-d252-4072-aded-c1166beefd89" (UID: "8647a121-d252-4072-aded-c1166beefd89"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.137668 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8647a121-d252-4072-aded-c1166beefd89" (UID: "8647a121-d252-4072-aded-c1166beefd89"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.203789 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fs4qh\" (UniqueName: \"kubernetes.io/projected/8647a121-d252-4072-aded-c1166beefd89-kube-api-access-fs4qh\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.203831 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.203846 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.203859 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8647a121-d252-4072-aded-c1166beefd89-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.494478 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" event={"ID":"8647a121-d252-4072-aded-c1166beefd89","Type":"ContainerDied","Data":"c1826a691ee10c947f25d41225832bbbec7fe80bf9a84d983d16952aa5702d4d"} Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.494523 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1826a691ee10c947f25d41225832bbbec7fe80bf9a84d983d16952aa5702d4d" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.494614 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-openstack-openstack-cell1-nq7vm" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.589796 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-openstack-openstack-cell1-jdvdr"] Dec 04 13:46:41 crc kubenswrapper[4979]: E1204 13:46:41.590456 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8647a121-d252-4072-aded-c1166beefd89" containerName="validate-network-openstack-openstack-cell1" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.590503 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8647a121-d252-4072-aded-c1166beefd89" containerName="validate-network-openstack-openstack-cell1" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.590773 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8647a121-d252-4072-aded-c1166beefd89" containerName="validate-network-openstack-openstack-cell1" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.591770 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.593860 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.594283 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.594958 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.595276 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.602522 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-jdvdr"] Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.717895 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-inventory\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.718268 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ceph\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.718349 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ssh-key\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.718534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kk54\" (UniqueName: \"kubernetes.io/projected/7fed6f42-de65-4d92-9532-57de104f61de-kube-api-access-4kk54\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.820531 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-inventory\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.820602 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ceph\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.820675 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ssh-key\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.820789 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kk54\" (UniqueName: \"kubernetes.io/projected/7fed6f42-de65-4d92-9532-57de104f61de-kube-api-access-4kk54\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.827726 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ssh-key\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.828213 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-inventory\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.828383 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ceph\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.851133 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kk54\" (UniqueName: \"kubernetes.io/projected/7fed6f42-de65-4d92-9532-57de104f61de-kube-api-access-4kk54\") pod \"install-os-openstack-openstack-cell1-jdvdr\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:41 crc kubenswrapper[4979]: I1204 13:46:41.911770 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:46:42 crc kubenswrapper[4979]: I1204 13:46:42.457432 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-openstack-openstack-cell1-jdvdr"] Dec 04 13:46:42 crc kubenswrapper[4979]: I1204 13:46:42.504148 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" event={"ID":"7fed6f42-de65-4d92-9532-57de104f61de","Type":"ContainerStarted","Data":"2d1fbd081c63d7ba836867da1f47bedb6fce69da32f911ee689c3812bdf8c964"} Dec 04 13:46:44 crc kubenswrapper[4979]: I1204 13:46:44.525192 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" event={"ID":"7fed6f42-de65-4d92-9532-57de104f61de","Type":"ContainerStarted","Data":"4feb43d064bff577007d86f3a56f741a83d1eb39f2fe076a7905788befab8a37"} Dec 04 13:47:27 crc kubenswrapper[4979]: I1204 13:47:27.001025 4979 generic.go:334] "Generic (PLEG): container finished" podID="7fed6f42-de65-4d92-9532-57de104f61de" containerID="4feb43d064bff577007d86f3a56f741a83d1eb39f2fe076a7905788befab8a37" exitCode=0 Dec 04 13:47:27 crc kubenswrapper[4979]: I1204 13:47:27.001056 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" event={"ID":"7fed6f42-de65-4d92-9532-57de104f61de","Type":"ContainerDied","Data":"4feb43d064bff577007d86f3a56f741a83d1eb39f2fe076a7905788befab8a37"} Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.445415 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.498118 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ssh-key\") pod \"7fed6f42-de65-4d92-9532-57de104f61de\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.498339 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-inventory\") pod \"7fed6f42-de65-4d92-9532-57de104f61de\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.498847 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kk54\" (UniqueName: \"kubernetes.io/projected/7fed6f42-de65-4d92-9532-57de104f61de-kube-api-access-4kk54\") pod \"7fed6f42-de65-4d92-9532-57de104f61de\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.498886 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ceph\") pod \"7fed6f42-de65-4d92-9532-57de104f61de\" (UID: \"7fed6f42-de65-4d92-9532-57de104f61de\") " Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.505849 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ceph" (OuterVolumeSpecName: "ceph") pod "7fed6f42-de65-4d92-9532-57de104f61de" (UID: "7fed6f42-de65-4d92-9532-57de104f61de"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.511785 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fed6f42-de65-4d92-9532-57de104f61de-kube-api-access-4kk54" (OuterVolumeSpecName: "kube-api-access-4kk54") pod "7fed6f42-de65-4d92-9532-57de104f61de" (UID: "7fed6f42-de65-4d92-9532-57de104f61de"). InnerVolumeSpecName "kube-api-access-4kk54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.534485 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-inventory" (OuterVolumeSpecName: "inventory") pod "7fed6f42-de65-4d92-9532-57de104f61de" (UID: "7fed6f42-de65-4d92-9532-57de104f61de"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.536217 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7fed6f42-de65-4d92-9532-57de104f61de" (UID: "7fed6f42-de65-4d92-9532-57de104f61de"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.603711 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.604682 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kk54\" (UniqueName: \"kubernetes.io/projected/7fed6f42-de65-4d92-9532-57de104f61de-kube-api-access-4kk54\") on node \"crc\" DevicePath \"\"" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.604729 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:47:28 crc kubenswrapper[4979]: I1204 13:47:28.604745 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7fed6f42-de65-4d92-9532-57de104f61de-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.024870 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" event={"ID":"7fed6f42-de65-4d92-9532-57de104f61de","Type":"ContainerDied","Data":"2d1fbd081c63d7ba836867da1f47bedb6fce69da32f911ee689c3812bdf8c964"} Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.024908 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d1fbd081c63d7ba836867da1f47bedb6fce69da32f911ee689c3812bdf8c964" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.024924 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-openstack-openstack-cell1-jdvdr" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.117868 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-z8bpb"] Dec 04 13:47:29 crc kubenswrapper[4979]: E1204 13:47:29.119236 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fed6f42-de65-4d92-9532-57de104f61de" containerName="install-os-openstack-openstack-cell1" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.119394 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fed6f42-de65-4d92-9532-57de104f61de" containerName="install-os-openstack-openstack-cell1" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.119831 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fed6f42-de65-4d92-9532-57de104f61de" containerName="install-os-openstack-openstack-cell1" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.121391 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.125145 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.125511 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.125677 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.125996 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.130082 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-z8bpb"] Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.221563 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br6rs\" (UniqueName: \"kubernetes.io/projected/35f1f69f-68fd-4931-84ae-52547e563860-kube-api-access-br6rs\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.221602 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ssh-key\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.221665 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-inventory\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.221685 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ceph\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.325039 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br6rs\" (UniqueName: \"kubernetes.io/projected/35f1f69f-68fd-4931-84ae-52547e563860-kube-api-access-br6rs\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.325151 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ssh-key\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.325430 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-inventory\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.325499 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ceph\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.332313 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-inventory\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.332490 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ssh-key\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.332858 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ceph\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.346073 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br6rs\" (UniqueName: \"kubernetes.io/projected/35f1f69f-68fd-4931-84ae-52547e563860-kube-api-access-br6rs\") pod \"configure-os-openstack-openstack-cell1-z8bpb\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:29 crc kubenswrapper[4979]: I1204 13:47:29.456510 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:47:30 crc kubenswrapper[4979]: I1204 13:47:30.011359 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-openstack-openstack-cell1-z8bpb"] Dec 04 13:47:30 crc kubenswrapper[4979]: I1204 13:47:30.039657 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" event={"ID":"35f1f69f-68fd-4931-84ae-52547e563860","Type":"ContainerStarted","Data":"81ae38af0e45ff5cbafb92f3daef327118cf62b567a8e52f574382f29b06a07d"} Dec 04 13:47:33 crc kubenswrapper[4979]: I1204 13:47:33.096909 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" event={"ID":"35f1f69f-68fd-4931-84ae-52547e563860","Type":"ContainerStarted","Data":"a46be7c2ed246731466d1253b8ae946e6400a2476bc12316369b1b4f3573918a"} Dec 04 13:47:33 crc kubenswrapper[4979]: I1204 13:47:33.117053 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" podStartSLOduration=2.338978056 podStartE2EDuration="4.117034041s" podCreationTimestamp="2025-12-04 13:47:29 +0000 UTC" firstStartedPulling="2025-12-04 13:47:30.017142876 +0000 UTC m=+7474.291438680" lastFinishedPulling="2025-12-04 13:47:31.795198841 +0000 UTC m=+7476.069494665" observedRunningTime="2025-12-04 13:47:33.112957842 +0000 UTC m=+7477.387253646" watchObservedRunningTime="2025-12-04 13:47:33.117034041 +0000 UTC m=+7477.391329845" Dec 04 13:48:16 crc kubenswrapper[4979]: I1204 13:48:16.511238 4979 generic.go:334] "Generic (PLEG): container finished" podID="35f1f69f-68fd-4931-84ae-52547e563860" containerID="a46be7c2ed246731466d1253b8ae946e6400a2476bc12316369b1b4f3573918a" exitCode=0 Dec 04 13:48:16 crc kubenswrapper[4979]: I1204 13:48:16.511450 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" event={"ID":"35f1f69f-68fd-4931-84ae-52547e563860","Type":"ContainerDied","Data":"a46be7c2ed246731466d1253b8ae946e6400a2476bc12316369b1b4f3573918a"} Dec 04 13:48:17 crc kubenswrapper[4979]: I1204 13:48:17.987855 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.060644 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ssh-key\") pod \"35f1f69f-68fd-4931-84ae-52547e563860\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.060823 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-inventory\") pod \"35f1f69f-68fd-4931-84ae-52547e563860\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.061087 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br6rs\" (UniqueName: \"kubernetes.io/projected/35f1f69f-68fd-4931-84ae-52547e563860-kube-api-access-br6rs\") pod \"35f1f69f-68fd-4931-84ae-52547e563860\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.061152 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ceph\") pod \"35f1f69f-68fd-4931-84ae-52547e563860\" (UID: \"35f1f69f-68fd-4931-84ae-52547e563860\") " Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.066560 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35f1f69f-68fd-4931-84ae-52547e563860-kube-api-access-br6rs" (OuterVolumeSpecName: "kube-api-access-br6rs") pod "35f1f69f-68fd-4931-84ae-52547e563860" (UID: "35f1f69f-68fd-4931-84ae-52547e563860"). InnerVolumeSpecName "kube-api-access-br6rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.066901 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ceph" (OuterVolumeSpecName: "ceph") pod "35f1f69f-68fd-4931-84ae-52547e563860" (UID: "35f1f69f-68fd-4931-84ae-52547e563860"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.090937 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "35f1f69f-68fd-4931-84ae-52547e563860" (UID: "35f1f69f-68fd-4931-84ae-52547e563860"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.093833 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-inventory" (OuterVolumeSpecName: "inventory") pod "35f1f69f-68fd-4931-84ae-52547e563860" (UID: "35f1f69f-68fd-4931-84ae-52547e563860"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.163212 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br6rs\" (UniqueName: \"kubernetes.io/projected/35f1f69f-68fd-4931-84ae-52547e563860-kube-api-access-br6rs\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.163244 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.163254 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.163263 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/35f1f69f-68fd-4931-84ae-52547e563860-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.530466 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" event={"ID":"35f1f69f-68fd-4931-84ae-52547e563860","Type":"ContainerDied","Data":"81ae38af0e45ff5cbafb92f3daef327118cf62b567a8e52f574382f29b06a07d"} Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.530854 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81ae38af0e45ff5cbafb92f3daef327118cf62b567a8e52f574382f29b06a07d" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.530516 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-openstack-openstack-cell1-z8bpb" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.639160 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-openstack-5rvvt"] Dec 04 13:48:18 crc kubenswrapper[4979]: E1204 13:48:18.639743 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35f1f69f-68fd-4931-84ae-52547e563860" containerName="configure-os-openstack-openstack-cell1" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.639771 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="35f1f69f-68fd-4931-84ae-52547e563860" containerName="configure-os-openstack-openstack-cell1" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.640180 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="35f1f69f-68fd-4931-84ae-52547e563860" containerName="configure-os-openstack-openstack-cell1" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.641463 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.644723 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.644830 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.644857 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.645105 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.675007 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6t29c\" (UniqueName: \"kubernetes.io/projected/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-kube-api-access-6t29c\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.675216 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ceph\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.675369 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-inventory-0\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.675408 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.693147 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-5rvvt"] Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.777462 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-inventory-0\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.777522 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.777564 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6t29c\" (UniqueName: \"kubernetes.io/projected/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-kube-api-access-6t29c\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.777694 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ceph\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.782889 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ceph\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.782930 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-inventory-0\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.790710 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ssh-key-openstack-cell1\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.793504 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6t29c\" (UniqueName: \"kubernetes.io/projected/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-kube-api-access-6t29c\") pod \"ssh-known-hosts-openstack-5rvvt\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:18 crc kubenswrapper[4979]: I1204 13:48:18.974466 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:19 crc kubenswrapper[4979]: I1204 13:48:19.531340 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-openstack-5rvvt"] Dec 04 13:48:19 crc kubenswrapper[4979]: W1204 13:48:19.536122 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63ae0b49_5321_4eff_96c5_3c3e22a1ea8d.slice/crio-512d6c2bf68c91034cecbf11fe17ac8296166f003301b78ac95386d60c1b0821 WatchSource:0}: Error finding container 512d6c2bf68c91034cecbf11fe17ac8296166f003301b78ac95386d60c1b0821: Status 404 returned error can't find the container with id 512d6c2bf68c91034cecbf11fe17ac8296166f003301b78ac95386d60c1b0821 Dec 04 13:48:20 crc kubenswrapper[4979]: I1204 13:48:20.554473 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-5rvvt" event={"ID":"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d","Type":"ContainerStarted","Data":"5306b0892fb4db15a09b3c1a1195f656135923881ede9e4223c355a87539680c"} Dec 04 13:48:20 crc kubenswrapper[4979]: I1204 13:48:20.554754 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-5rvvt" event={"ID":"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d","Type":"ContainerStarted","Data":"512d6c2bf68c91034cecbf11fe17ac8296166f003301b78ac95386d60c1b0821"} Dec 04 13:48:20 crc kubenswrapper[4979]: I1204 13:48:20.576913 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-openstack-5rvvt" podStartSLOduration=2.180250908 podStartE2EDuration="2.576890445s" podCreationTimestamp="2025-12-04 13:48:18 +0000 UTC" firstStartedPulling="2025-12-04 13:48:19.565534192 +0000 UTC m=+7523.839829996" lastFinishedPulling="2025-12-04 13:48:19.962173729 +0000 UTC m=+7524.236469533" observedRunningTime="2025-12-04 13:48:20.571177652 +0000 UTC m=+7524.845473466" watchObservedRunningTime="2025-12-04 13:48:20.576890445 +0000 UTC m=+7524.851186259" Dec 04 13:48:28 crc kubenswrapper[4979]: I1204 13:48:28.041644 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:48:28 crc kubenswrapper[4979]: I1204 13:48:28.042680 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:48:28 crc kubenswrapper[4979]: I1204 13:48:28.629675 4979 generic.go:334] "Generic (PLEG): container finished" podID="63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" containerID="5306b0892fb4db15a09b3c1a1195f656135923881ede9e4223c355a87539680c" exitCode=0 Dec 04 13:48:28 crc kubenswrapper[4979]: I1204 13:48:28.629754 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-5rvvt" event={"ID":"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d","Type":"ContainerDied","Data":"5306b0892fb4db15a09b3c1a1195f656135923881ede9e4223c355a87539680c"} Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.151894 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.247953 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6t29c\" (UniqueName: \"kubernetes.io/projected/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-kube-api-access-6t29c\") pod \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.248070 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ceph\") pod \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.248104 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-inventory-0\") pod \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.248452 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ssh-key-openstack-cell1\") pod \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\" (UID: \"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d\") " Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.254502 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ceph" (OuterVolumeSpecName: "ceph") pod "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" (UID: "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.256691 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-kube-api-access-6t29c" (OuterVolumeSpecName: "kube-api-access-6t29c") pod "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" (UID: "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d"). InnerVolumeSpecName "kube-api-access-6t29c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.278188 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" (UID: "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.303525 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ssh-key-openstack-cell1" (OuterVolumeSpecName: "ssh-key-openstack-cell1") pod "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" (UID: "63ae0b49-5321-4eff-96c5-3c3e22a1ea8d"). InnerVolumeSpecName "ssh-key-openstack-cell1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.351515 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-cell1\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ssh-key-openstack-cell1\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.351550 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6t29c\" (UniqueName: \"kubernetes.io/projected/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-kube-api-access-6t29c\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.351562 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.351576 4979 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/63ae0b49-5321-4eff-96c5-3c3e22a1ea8d-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.653036 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-openstack-5rvvt" event={"ID":"63ae0b49-5321-4eff-96c5-3c3e22a1ea8d","Type":"ContainerDied","Data":"512d6c2bf68c91034cecbf11fe17ac8296166f003301b78ac95386d60c1b0821"} Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.653078 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="512d6c2bf68c91034cecbf11fe17ac8296166f003301b78ac95386d60c1b0821" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.653139 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-openstack-5rvvt" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.731462 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-openstack-openstack-cell1-dncws"] Dec 04 13:48:30 crc kubenswrapper[4979]: E1204 13:48:30.732178 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" containerName="ssh-known-hosts-openstack" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.732209 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" containerName="ssh-known-hosts-openstack" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.732641 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ae0b49-5321-4eff-96c5-3c3e22a1ea8d" containerName="ssh-known-hosts-openstack" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.733845 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.736278 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.736435 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.736592 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.736589 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.764433 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-dncws"] Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.862576 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-inventory\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.862687 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ceph\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.863005 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf5dg\" (UniqueName: \"kubernetes.io/projected/c75e0fbc-b974-4306-83b7-20c82ac841e7-kube-api-access-bf5dg\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.863521 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ssh-key\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.965646 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ceph\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.965801 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf5dg\" (UniqueName: \"kubernetes.io/projected/c75e0fbc-b974-4306-83b7-20c82ac841e7-kube-api-access-bf5dg\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.965878 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ssh-key\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.965927 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-inventory\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.971371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ssh-key\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.971371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-inventory\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.971413 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ceph\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:30 crc kubenswrapper[4979]: I1204 13:48:30.985207 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf5dg\" (UniqueName: \"kubernetes.io/projected/c75e0fbc-b974-4306-83b7-20c82ac841e7-kube-api-access-bf5dg\") pod \"run-os-openstack-openstack-cell1-dncws\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:31 crc kubenswrapper[4979]: I1204 13:48:31.060070 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:31 crc kubenswrapper[4979]: I1204 13:48:31.632535 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-openstack-openstack-cell1-dncws"] Dec 04 13:48:31 crc kubenswrapper[4979]: I1204 13:48:31.640178 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:48:31 crc kubenswrapper[4979]: I1204 13:48:31.664499 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dncws" event={"ID":"c75e0fbc-b974-4306-83b7-20c82ac841e7","Type":"ContainerStarted","Data":"718addc079023b57e84281bd0f45877c8de31dbc59b4c7580ab8dde0002db3c6"} Dec 04 13:48:32 crc kubenswrapper[4979]: I1204 13:48:32.678487 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dncws" event={"ID":"c75e0fbc-b974-4306-83b7-20c82ac841e7","Type":"ContainerStarted","Data":"9bed96a8147a5927888bbfe5e27320b53a3b8d4a2d2f169b205d97bf63550981"} Dec 04 13:48:32 crc kubenswrapper[4979]: I1204 13:48:32.707561 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-openstack-openstack-cell1-dncws" podStartSLOduration=2.009391841 podStartE2EDuration="2.707540904s" podCreationTimestamp="2025-12-04 13:48:30 +0000 UTC" firstStartedPulling="2025-12-04 13:48:31.639914852 +0000 UTC m=+7535.914210656" lastFinishedPulling="2025-12-04 13:48:32.338063915 +0000 UTC m=+7536.612359719" observedRunningTime="2025-12-04 13:48:32.699829407 +0000 UTC m=+7536.974125221" watchObservedRunningTime="2025-12-04 13:48:32.707540904 +0000 UTC m=+7536.981836708" Dec 04 13:48:41 crc kubenswrapper[4979]: I1204 13:48:41.786402 4979 generic.go:334] "Generic (PLEG): container finished" podID="c75e0fbc-b974-4306-83b7-20c82ac841e7" containerID="9bed96a8147a5927888bbfe5e27320b53a3b8d4a2d2f169b205d97bf63550981" exitCode=0 Dec 04 13:48:41 crc kubenswrapper[4979]: I1204 13:48:41.786526 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dncws" event={"ID":"c75e0fbc-b974-4306-83b7-20c82ac841e7","Type":"ContainerDied","Data":"9bed96a8147a5927888bbfe5e27320b53a3b8d4a2d2f169b205d97bf63550981"} Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.269331 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.360713 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-inventory\") pod \"c75e0fbc-b974-4306-83b7-20c82ac841e7\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.361071 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ssh-key\") pod \"c75e0fbc-b974-4306-83b7-20c82ac841e7\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.361107 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf5dg\" (UniqueName: \"kubernetes.io/projected/c75e0fbc-b974-4306-83b7-20c82ac841e7-kube-api-access-bf5dg\") pod \"c75e0fbc-b974-4306-83b7-20c82ac841e7\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.361157 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ceph\") pod \"c75e0fbc-b974-4306-83b7-20c82ac841e7\" (UID: \"c75e0fbc-b974-4306-83b7-20c82ac841e7\") " Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.387707 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ceph" (OuterVolumeSpecName: "ceph") pod "c75e0fbc-b974-4306-83b7-20c82ac841e7" (UID: "c75e0fbc-b974-4306-83b7-20c82ac841e7"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.388210 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c75e0fbc-b974-4306-83b7-20c82ac841e7-kube-api-access-bf5dg" (OuterVolumeSpecName: "kube-api-access-bf5dg") pod "c75e0fbc-b974-4306-83b7-20c82ac841e7" (UID: "c75e0fbc-b974-4306-83b7-20c82ac841e7"). InnerVolumeSpecName "kube-api-access-bf5dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.391207 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-inventory" (OuterVolumeSpecName: "inventory") pod "c75e0fbc-b974-4306-83b7-20c82ac841e7" (UID: "c75e0fbc-b974-4306-83b7-20c82ac841e7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.391268 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c75e0fbc-b974-4306-83b7-20c82ac841e7" (UID: "c75e0fbc-b974-4306-83b7-20c82ac841e7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.463867 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf5dg\" (UniqueName: \"kubernetes.io/projected/c75e0fbc-b974-4306-83b7-20c82ac841e7-kube-api-access-bf5dg\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.463915 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.463936 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.463951 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c75e0fbc-b974-4306-83b7-20c82ac841e7-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.806855 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-openstack-openstack-cell1-dncws" event={"ID":"c75e0fbc-b974-4306-83b7-20c82ac841e7","Type":"ContainerDied","Data":"718addc079023b57e84281bd0f45877c8de31dbc59b4c7580ab8dde0002db3c6"} Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.806901 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="718addc079023b57e84281bd0f45877c8de31dbc59b4c7580ab8dde0002db3c6" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.806913 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-openstack-openstack-cell1-dncws" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.904355 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-r6nzl"] Dec 04 13:48:43 crc kubenswrapper[4979]: E1204 13:48:43.904968 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c75e0fbc-b974-4306-83b7-20c82ac841e7" containerName="run-os-openstack-openstack-cell1" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.904983 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="c75e0fbc-b974-4306-83b7-20c82ac841e7" containerName="run-os-openstack-openstack-cell1" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.905229 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="c75e0fbc-b974-4306-83b7-20c82ac841e7" containerName="run-os-openstack-openstack-cell1" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.906988 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.910627 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.910841 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.911095 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.914884 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:48:43 crc kubenswrapper[4979]: E1204 13:48:43.938585 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc75e0fbc_b974_4306_83b7_20c82ac841e7.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc75e0fbc_b974_4306_83b7_20c82ac841e7.slice/crio-718addc079023b57e84281bd0f45877c8de31dbc59b4c7580ab8dde0002db3c6\": RecentStats: unable to find data in memory cache]" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.961790 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-r6nzl"] Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.977250 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-inventory\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.977596 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.977739 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ngjb\" (UniqueName: \"kubernetes.io/projected/ff6b5d75-375e-4c64-a827-089952fb1d92-kube-api-access-5ngjb\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:43 crc kubenswrapper[4979]: I1204 13:48:43.977943 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ceph\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.079459 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-inventory\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.079598 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.079641 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ngjb\" (UniqueName: \"kubernetes.io/projected/ff6b5d75-375e-4c64-a827-089952fb1d92-kube-api-access-5ngjb\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.079709 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ceph\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.087138 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ceph\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.090700 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ssh-key\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.091805 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-inventory\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.113152 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ngjb\" (UniqueName: \"kubernetes.io/projected/ff6b5d75-375e-4c64-a827-089952fb1d92-kube-api-access-5ngjb\") pod \"reboot-os-openstack-openstack-cell1-r6nzl\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.281954 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:48:44 crc kubenswrapper[4979]: W1204 13:48:44.861878 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff6b5d75_375e_4c64_a827_089952fb1d92.slice/crio-9e0fa00cd5a1d6e41cdc6e375a83f7268dcb5514fb788203cb0fd15e35a6f611 WatchSource:0}: Error finding container 9e0fa00cd5a1d6e41cdc6e375a83f7268dcb5514fb788203cb0fd15e35a6f611: Status 404 returned error can't find the container with id 9e0fa00cd5a1d6e41cdc6e375a83f7268dcb5514fb788203cb0fd15e35a6f611 Dec 04 13:48:44 crc kubenswrapper[4979]: I1204 13:48:44.864185 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-openstack-openstack-cell1-r6nzl"] Dec 04 13:48:45 crc kubenswrapper[4979]: I1204 13:48:45.826963 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" event={"ID":"ff6b5d75-375e-4c64-a827-089952fb1d92","Type":"ContainerStarted","Data":"9ceb9b3c873b46770fc15edd5f5c9616171b22b4237953e0aed503dee816f8a7"} Dec 04 13:48:45 crc kubenswrapper[4979]: I1204 13:48:45.828437 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" event={"ID":"ff6b5d75-375e-4c64-a827-089952fb1d92","Type":"ContainerStarted","Data":"9e0fa00cd5a1d6e41cdc6e375a83f7268dcb5514fb788203cb0fd15e35a6f611"} Dec 04 13:48:45 crc kubenswrapper[4979]: I1204 13:48:45.844789 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" podStartSLOduration=2.459237828 podStartE2EDuration="2.844772478s" podCreationTimestamp="2025-12-04 13:48:43 +0000 UTC" firstStartedPulling="2025-12-04 13:48:44.865450434 +0000 UTC m=+7549.139746238" lastFinishedPulling="2025-12-04 13:48:45.250985084 +0000 UTC m=+7549.525280888" observedRunningTime="2025-12-04 13:48:45.842705963 +0000 UTC m=+7550.117001767" watchObservedRunningTime="2025-12-04 13:48:45.844772478 +0000 UTC m=+7550.119068282" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.561836 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2qxsp"] Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.566390 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.584583 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2qxsp"] Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.696184 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tnm9\" (UniqueName: \"kubernetes.io/projected/17b27914-8c7c-4f90-a9b9-f061ee34fc60-kube-api-access-7tnm9\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.696282 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-utilities\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.696394 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-catalog-content\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.797814 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-catalog-content\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.797935 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tnm9\" (UniqueName: \"kubernetes.io/projected/17b27914-8c7c-4f90-a9b9-f061ee34fc60-kube-api-access-7tnm9\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.798016 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-utilities\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.798475 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-catalog-content\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.798531 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-utilities\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.830819 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tnm9\" (UniqueName: \"kubernetes.io/projected/17b27914-8c7c-4f90-a9b9-f061ee34fc60-kube-api-access-7tnm9\") pod \"certified-operators-2qxsp\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:55 crc kubenswrapper[4979]: I1204 13:48:55.895866 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:48:56 crc kubenswrapper[4979]: I1204 13:48:56.475525 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2qxsp"] Dec 04 13:48:56 crc kubenswrapper[4979]: W1204 13:48:56.501331 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17b27914_8c7c_4f90_a9b9_f061ee34fc60.slice/crio-c14f9ac0dcf18c9378ca62472fbce97beb094fde9af8c88b926b77821efcc39b WatchSource:0}: Error finding container c14f9ac0dcf18c9378ca62472fbce97beb094fde9af8c88b926b77821efcc39b: Status 404 returned error can't find the container with id c14f9ac0dcf18c9378ca62472fbce97beb094fde9af8c88b926b77821efcc39b Dec 04 13:48:56 crc kubenswrapper[4979]: I1204 13:48:56.942641 4979 generic.go:334] "Generic (PLEG): container finished" podID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerID="871f573f5bbc1984ccae5ee2eb1a177d1167883007d812493cd40e2cfdcf46d8" exitCode=0 Dec 04 13:48:56 crc kubenswrapper[4979]: I1204 13:48:56.943000 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerDied","Data":"871f573f5bbc1984ccae5ee2eb1a177d1167883007d812493cd40e2cfdcf46d8"} Dec 04 13:48:56 crc kubenswrapper[4979]: I1204 13:48:56.943454 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerStarted","Data":"c14f9ac0dcf18c9378ca62472fbce97beb094fde9af8c88b926b77821efcc39b"} Dec 04 13:48:57 crc kubenswrapper[4979]: I1204 13:48:57.960896 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerStarted","Data":"6f320ae6dcb891181689d6e2eb8b3697c45d9e01f4b7649cb6d45cad4bce566a"} Dec 04 13:48:58 crc kubenswrapper[4979]: I1204 13:48:58.040892 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:48:58 crc kubenswrapper[4979]: I1204 13:48:58.040967 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:48:58 crc kubenswrapper[4979]: I1204 13:48:58.974171 4979 generic.go:334] "Generic (PLEG): container finished" podID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerID="6f320ae6dcb891181689d6e2eb8b3697c45d9e01f4b7649cb6d45cad4bce566a" exitCode=0 Dec 04 13:48:58 crc kubenswrapper[4979]: I1204 13:48:58.974218 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerDied","Data":"6f320ae6dcb891181689d6e2eb8b3697c45d9e01f4b7649cb6d45cad4bce566a"} Dec 04 13:48:59 crc kubenswrapper[4979]: I1204 13:48:59.991501 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerStarted","Data":"4ab9cbf787ecbd532648af29f37046c2af78fc369a47e57c0de39ec8ab3c3baf"} Dec 04 13:49:00 crc kubenswrapper[4979]: I1204 13:49:00.012125 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2qxsp" podStartSLOduration=2.226858402 podStartE2EDuration="5.012105509s" podCreationTimestamp="2025-12-04 13:48:55 +0000 UTC" firstStartedPulling="2025-12-04 13:48:56.947028987 +0000 UTC m=+7561.221324791" lastFinishedPulling="2025-12-04 13:48:59.732276094 +0000 UTC m=+7564.006571898" observedRunningTime="2025-12-04 13:49:00.007452683 +0000 UTC m=+7564.281748497" watchObservedRunningTime="2025-12-04 13:49:00.012105509 +0000 UTC m=+7564.286401313" Dec 04 13:49:02 crc kubenswrapper[4979]: I1204 13:49:02.014026 4979 generic.go:334] "Generic (PLEG): container finished" podID="ff6b5d75-375e-4c64-a827-089952fb1d92" containerID="9ceb9b3c873b46770fc15edd5f5c9616171b22b4237953e0aed503dee816f8a7" exitCode=0 Dec 04 13:49:02 crc kubenswrapper[4979]: I1204 13:49:02.014121 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" event={"ID":"ff6b5d75-375e-4c64-a827-089952fb1d92","Type":"ContainerDied","Data":"9ceb9b3c873b46770fc15edd5f5c9616171b22b4237953e0aed503dee816f8a7"} Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.691506 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.776291 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-inventory\") pod \"ff6b5d75-375e-4c64-a827-089952fb1d92\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.776772 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ceph\") pod \"ff6b5d75-375e-4c64-a827-089952fb1d92\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.776924 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ngjb\" (UniqueName: \"kubernetes.io/projected/ff6b5d75-375e-4c64-a827-089952fb1d92-kube-api-access-5ngjb\") pod \"ff6b5d75-375e-4c64-a827-089952fb1d92\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.777060 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ssh-key\") pod \"ff6b5d75-375e-4c64-a827-089952fb1d92\" (UID: \"ff6b5d75-375e-4c64-a827-089952fb1d92\") " Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.788208 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff6b5d75-375e-4c64-a827-089952fb1d92-kube-api-access-5ngjb" (OuterVolumeSpecName: "kube-api-access-5ngjb") pod "ff6b5d75-375e-4c64-a827-089952fb1d92" (UID: "ff6b5d75-375e-4c64-a827-089952fb1d92"). InnerVolumeSpecName "kube-api-access-5ngjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.789189 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ceph" (OuterVolumeSpecName: "ceph") pod "ff6b5d75-375e-4c64-a827-089952fb1d92" (UID: "ff6b5d75-375e-4c64-a827-089952fb1d92"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.811719 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-inventory" (OuterVolumeSpecName: "inventory") pod "ff6b5d75-375e-4c64-a827-089952fb1d92" (UID: "ff6b5d75-375e-4c64-a827-089952fb1d92"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.814658 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ff6b5d75-375e-4c64-a827-089952fb1d92" (UID: "ff6b5d75-375e-4c64-a827-089952fb1d92"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.880390 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.880427 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.880438 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ff6b5d75-375e-4c64-a827-089952fb1d92-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:03 crc kubenswrapper[4979]: I1204 13:49:03.880453 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ngjb\" (UniqueName: \"kubernetes.io/projected/ff6b5d75-375e-4c64-a827-089952fb1d92-kube-api-access-5ngjb\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.035877 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" event={"ID":"ff6b5d75-375e-4c64-a827-089952fb1d92","Type":"ContainerDied","Data":"9e0fa00cd5a1d6e41cdc6e375a83f7268dcb5514fb788203cb0fd15e35a6f611"} Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.035930 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-openstack-openstack-cell1-r6nzl" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.035939 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e0fa00cd5a1d6e41cdc6e375a83f7268dcb5514fb788203cb0fd15e35a6f611" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.497838 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vlmvv"] Dec 04 13:49:04 crc kubenswrapper[4979]: E1204 13:49:04.500072 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff6b5d75-375e-4c64-a827-089952fb1d92" containerName="reboot-os-openstack-openstack-cell1" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.500115 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff6b5d75-375e-4c64-a827-089952fb1d92" containerName="reboot-os-openstack-openstack-cell1" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.500907 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff6b5d75-375e-4c64-a827-089952fb1d92" containerName="reboot-os-openstack-openstack-cell1" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.502279 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.506270 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.506467 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.509338 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.509665 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.535060 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vlmvv"] Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.605717 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.605955 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.605995 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-inventory\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606162 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ceph\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606280 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606381 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606512 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606571 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606625 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606648 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606681 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhwk5\" (UniqueName: \"kubernetes.io/projected/2fc0f4a7-a63e-416b-b4da-b5f488c06289-kube-api-access-hhwk5\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.606762 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709170 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709257 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709274 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709313 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhwk5\" (UniqueName: \"kubernetes.io/projected/2fc0f4a7-a63e-416b-b4da-b5f488c06289-kube-api-access-hhwk5\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709351 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709373 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709414 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709430 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-inventory\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709484 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ceph\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709533 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.709578 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.714483 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-inventory\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.715407 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ssh-key\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.715980 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ceph\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.718000 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ovn-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.718126 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-libvirt-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.718767 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-metadata-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.718825 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-sriov-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.720964 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-nova-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.721021 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-telemetry-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.722484 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-dhcp-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.723801 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-bootstrap-combined-ca-bundle\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.725645 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhwk5\" (UniqueName: \"kubernetes.io/projected/2fc0f4a7-a63e-416b-b4da-b5f488c06289-kube-api-access-hhwk5\") pod \"install-certs-openstack-openstack-cell1-vlmvv\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:04 crc kubenswrapper[4979]: I1204 13:49:04.832513 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:05 crc kubenswrapper[4979]: I1204 13:49:05.421179 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-openstack-openstack-cell1-vlmvv"] Dec 04 13:49:05 crc kubenswrapper[4979]: I1204 13:49:05.899962 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:49:05 crc kubenswrapper[4979]: I1204 13:49:05.900384 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:49:05 crc kubenswrapper[4979]: I1204 13:49:05.971225 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:49:06 crc kubenswrapper[4979]: I1204 13:49:06.057408 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" event={"ID":"2fc0f4a7-a63e-416b-b4da-b5f488c06289","Type":"ContainerStarted","Data":"e34dfcf847ddee6f9fb8c058ffb41245e6e3eba635f3ac21a9de015ec082132a"} Dec 04 13:49:06 crc kubenswrapper[4979]: I1204 13:49:06.104995 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:49:06 crc kubenswrapper[4979]: I1204 13:49:06.211230 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2qxsp"] Dec 04 13:49:07 crc kubenswrapper[4979]: I1204 13:49:07.077555 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" event={"ID":"2fc0f4a7-a63e-416b-b4da-b5f488c06289","Type":"ContainerStarted","Data":"15a699eaaf0bd74d1185fa7da8cb2b30e2ca468918c2f728b3cbe31bb58f9cf7"} Dec 04 13:49:07 crc kubenswrapper[4979]: I1204 13:49:07.100845 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" podStartSLOduration=2.709153475 podStartE2EDuration="3.100807628s" podCreationTimestamp="2025-12-04 13:49:04 +0000 UTC" firstStartedPulling="2025-12-04 13:49:05.427999096 +0000 UTC m=+7569.702294900" lastFinishedPulling="2025-12-04 13:49:05.819653249 +0000 UTC m=+7570.093949053" observedRunningTime="2025-12-04 13:49:07.094257283 +0000 UTC m=+7571.368553097" watchObservedRunningTime="2025-12-04 13:49:07.100807628 +0000 UTC m=+7571.375103432" Dec 04 13:49:08 crc kubenswrapper[4979]: I1204 13:49:08.084848 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2qxsp" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="registry-server" containerID="cri-o://4ab9cbf787ecbd532648af29f37046c2af78fc369a47e57c0de39ec8ab3c3baf" gracePeriod=2 Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.103865 4979 generic.go:334] "Generic (PLEG): container finished" podID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerID="4ab9cbf787ecbd532648af29f37046c2af78fc369a47e57c0de39ec8ab3c3baf" exitCode=0 Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.103915 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerDied","Data":"4ab9cbf787ecbd532648af29f37046c2af78fc369a47e57c0de39ec8ab3c3baf"} Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.104185 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2qxsp" event={"ID":"17b27914-8c7c-4f90-a9b9-f061ee34fc60","Type":"ContainerDied","Data":"c14f9ac0dcf18c9378ca62472fbce97beb094fde9af8c88b926b77821efcc39b"} Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.104197 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c14f9ac0dcf18c9378ca62472fbce97beb094fde9af8c88b926b77821efcc39b" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.120716 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.211449 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tnm9\" (UniqueName: \"kubernetes.io/projected/17b27914-8c7c-4f90-a9b9-f061ee34fc60-kube-api-access-7tnm9\") pod \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.211580 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-utilities\") pod \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.212418 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-utilities" (OuterVolumeSpecName: "utilities") pod "17b27914-8c7c-4f90-a9b9-f061ee34fc60" (UID: "17b27914-8c7c-4f90-a9b9-f061ee34fc60"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.213052 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-catalog-content\") pod \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\" (UID: \"17b27914-8c7c-4f90-a9b9-f061ee34fc60\") " Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.216206 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.225208 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17b27914-8c7c-4f90-a9b9-f061ee34fc60-kube-api-access-7tnm9" (OuterVolumeSpecName: "kube-api-access-7tnm9") pod "17b27914-8c7c-4f90-a9b9-f061ee34fc60" (UID: "17b27914-8c7c-4f90-a9b9-f061ee34fc60"). InnerVolumeSpecName "kube-api-access-7tnm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.271749 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17b27914-8c7c-4f90-a9b9-f061ee34fc60" (UID: "17b27914-8c7c-4f90-a9b9-f061ee34fc60"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.318749 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17b27914-8c7c-4f90-a9b9-f061ee34fc60-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:09 crc kubenswrapper[4979]: I1204 13:49:09.318780 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tnm9\" (UniqueName: \"kubernetes.io/projected/17b27914-8c7c-4f90-a9b9-f061ee34fc60-kube-api-access-7tnm9\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:10 crc kubenswrapper[4979]: I1204 13:49:10.116014 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2qxsp" Dec 04 13:49:10 crc kubenswrapper[4979]: I1204 13:49:10.161952 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2qxsp"] Dec 04 13:49:10 crc kubenswrapper[4979]: I1204 13:49:10.172814 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2qxsp"] Dec 04 13:49:10 crc kubenswrapper[4979]: I1204 13:49:10.217989 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" path="/var/lib/kubelet/pods/17b27914-8c7c-4f90-a9b9-f061ee34fc60/volumes" Dec 04 13:49:26 crc kubenswrapper[4979]: I1204 13:49:26.285240 4979 generic.go:334] "Generic (PLEG): container finished" podID="2fc0f4a7-a63e-416b-b4da-b5f488c06289" containerID="15a699eaaf0bd74d1185fa7da8cb2b30e2ca468918c2f728b3cbe31bb58f9cf7" exitCode=0 Dec 04 13:49:26 crc kubenswrapper[4979]: I1204 13:49:26.285357 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" event={"ID":"2fc0f4a7-a63e-416b-b4da-b5f488c06289","Type":"ContainerDied","Data":"15a699eaaf0bd74d1185fa7da8cb2b30e2ca468918c2f728b3cbe31bb58f9cf7"} Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.865058 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939003 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-libvirt-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939053 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhwk5\" (UniqueName: \"kubernetes.io/projected/2fc0f4a7-a63e-416b-b4da-b5f488c06289-kube-api-access-hhwk5\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939131 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-bootstrap-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939167 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ceph\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939217 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ssh-key\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939275 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-inventory\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939342 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-dhcp-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939379 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-metadata-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939416 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-telemetry-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939457 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ovn-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939494 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-sriov-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.939540 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-nova-combined-ca-bundle\") pod \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\" (UID: \"2fc0f4a7-a63e-416b-b4da-b5f488c06289\") " Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.946869 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.946920 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.946994 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fc0f4a7-a63e-416b-b4da-b5f488c06289-kube-api-access-hhwk5" (OuterVolumeSpecName: "kube-api-access-hhwk5") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "kube-api-access-hhwk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.948369 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.949356 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.949418 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.949431 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.950898 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ceph" (OuterVolumeSpecName: "ceph") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.951534 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.954432 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.976715 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:27 crc kubenswrapper[4979]: I1204 13:49:27.982169 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-inventory" (OuterVolumeSpecName: "inventory") pod "2fc0f4a7-a63e-416b-b4da-b5f488c06289" (UID: "2fc0f4a7-a63e-416b-b4da-b5f488c06289"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.041576 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.041638 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.041707 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.042553 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4d83d72a3f1cfee91bd625d5ef34b5bb09b87f3b468acc46be5ce89bc808d8ac"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.042627 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://4d83d72a3f1cfee91bd625d5ef34b5bb09b87f3b468acc46be5ce89bc808d8ac" gracePeriod=600 Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.043970 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044021 4979 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044042 4979 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044060 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhwk5\" (UniqueName: \"kubernetes.io/projected/2fc0f4a7-a63e-416b-b4da-b5f488c06289-kube-api-access-hhwk5\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044079 4979 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044096 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044112 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044130 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044147 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044166 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044184 4979 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.044202 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2fc0f4a7-a63e-416b-b4da-b5f488c06289-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.308894 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" event={"ID":"2fc0f4a7-a63e-416b-b4da-b5f488c06289","Type":"ContainerDied","Data":"e34dfcf847ddee6f9fb8c058ffb41245e6e3eba635f3ac21a9de015ec082132a"} Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.309256 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e34dfcf847ddee6f9fb8c058ffb41245e6e3eba635f3ac21a9de015ec082132a" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.308909 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-openstack-openstack-cell1-vlmvv" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.313401 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="4d83d72a3f1cfee91bd625d5ef34b5bb09b87f3b468acc46be5ce89bc808d8ac" exitCode=0 Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.313446 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"4d83d72a3f1cfee91bd625d5ef34b5bb09b87f3b468acc46be5ce89bc808d8ac"} Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.313490 4979 scope.go:117] "RemoveContainer" containerID="41c09db5a67eb8220f0a14aafd557fd65f89614df62bc96bb7ed3f64fe02b15d" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387045 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-b5xn4"] Dec 04 13:49:28 crc kubenswrapper[4979]: E1204 13:49:28.387588 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="extract-content" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387610 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="extract-content" Dec 04 13:49:28 crc kubenswrapper[4979]: E1204 13:49:28.387627 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="extract-utilities" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387636 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="extract-utilities" Dec 04 13:49:28 crc kubenswrapper[4979]: E1204 13:49:28.387662 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fc0f4a7-a63e-416b-b4da-b5f488c06289" containerName="install-certs-openstack-openstack-cell1" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387670 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fc0f4a7-a63e-416b-b4da-b5f488c06289" containerName="install-certs-openstack-openstack-cell1" Dec 04 13:49:28 crc kubenswrapper[4979]: E1204 13:49:28.387690 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="registry-server" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387699 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="registry-server" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387954 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fc0f4a7-a63e-416b-b4da-b5f488c06289" containerName="install-certs-openstack-openstack-cell1" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.387979 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="17b27914-8c7c-4f90-a9b9-f061ee34fc60" containerName="registry-server" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.388885 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.395338 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.395583 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.395679 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.395894 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.398783 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-b5xn4"] Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.451650 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ceph\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.451924 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.452008 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pxb6\" (UniqueName: \"kubernetes.io/projected/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-kube-api-access-2pxb6\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.452086 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-inventory\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.554805 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ceph\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.555101 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.555201 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pxb6\" (UniqueName: \"kubernetes.io/projected/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-kube-api-access-2pxb6\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.555294 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-inventory\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.559707 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ceph\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.560120 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ssh-key\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.561983 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-inventory\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.576076 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pxb6\" (UniqueName: \"kubernetes.io/projected/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-kube-api-access-2pxb6\") pod \"ceph-client-openstack-openstack-cell1-b5xn4\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:28 crc kubenswrapper[4979]: I1204 13:49:28.717517 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:29 crc kubenswrapper[4979]: I1204 13:49:29.265337 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-openstack-openstack-cell1-b5xn4"] Dec 04 13:49:29 crc kubenswrapper[4979]: I1204 13:49:29.323741 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" event={"ID":"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba","Type":"ContainerStarted","Data":"0b4d2bb58e20d75d92f006fddb2527a31280c03fb89c8c3b80747fba3e9d2d4b"} Dec 04 13:49:29 crc kubenswrapper[4979]: I1204 13:49:29.326622 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3"} Dec 04 13:49:30 crc kubenswrapper[4979]: I1204 13:49:30.379716 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" event={"ID":"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba","Type":"ContainerStarted","Data":"b7d7ff0bbd5d637caee3a8488c4e0a91fb00a7da4d44b82e76a450fbf7f8484a"} Dec 04 13:49:30 crc kubenswrapper[4979]: I1204 13:49:30.412467 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" podStartSLOduration=1.9847372669999999 podStartE2EDuration="2.412446268s" podCreationTimestamp="2025-12-04 13:49:28 +0000 UTC" firstStartedPulling="2025-12-04 13:49:29.271851088 +0000 UTC m=+7593.546146892" lastFinishedPulling="2025-12-04 13:49:29.699560089 +0000 UTC m=+7593.973855893" observedRunningTime="2025-12-04 13:49:30.410694591 +0000 UTC m=+7594.684990395" watchObservedRunningTime="2025-12-04 13:49:30.412446268 +0000 UTC m=+7594.686742072" Dec 04 13:49:35 crc kubenswrapper[4979]: I1204 13:49:35.428889 4979 generic.go:334] "Generic (PLEG): container finished" podID="f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" containerID="b7d7ff0bbd5d637caee3a8488c4e0a91fb00a7da4d44b82e76a450fbf7f8484a" exitCode=0 Dec 04 13:49:35 crc kubenswrapper[4979]: I1204 13:49:35.428997 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" event={"ID":"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba","Type":"ContainerDied","Data":"b7d7ff0bbd5d637caee3a8488c4e0a91fb00a7da4d44b82e76a450fbf7f8484a"} Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.052898 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.182744 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ceph\") pod \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.182860 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ssh-key\") pod \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.182974 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pxb6\" (UniqueName: \"kubernetes.io/projected/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-kube-api-access-2pxb6\") pod \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.183012 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-inventory\") pod \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\" (UID: \"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba\") " Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.189120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ceph" (OuterVolumeSpecName: "ceph") pod "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" (UID: "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.190470 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-kube-api-access-2pxb6" (OuterVolumeSpecName: "kube-api-access-2pxb6") pod "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" (UID: "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba"). InnerVolumeSpecName "kube-api-access-2pxb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.217362 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-inventory" (OuterVolumeSpecName: "inventory") pod "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" (UID: "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.217418 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" (UID: "f1e47e68-9d7a-4202-97ff-eab1b0e2ccba"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.285638 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.285685 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.285696 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pxb6\" (UniqueName: \"kubernetes.io/projected/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-kube-api-access-2pxb6\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.285707 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f1e47e68-9d7a-4202-97ff-eab1b0e2ccba-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.479773 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" event={"ID":"f1e47e68-9d7a-4202-97ff-eab1b0e2ccba","Type":"ContainerDied","Data":"0b4d2bb58e20d75d92f006fddb2527a31280c03fb89c8c3b80747fba3e9d2d4b"} Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.479818 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b4d2bb58e20d75d92f006fddb2527a31280c03fb89c8c3b80747fba3e9d2d4b" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.479895 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-openstack-openstack-cell1-b5xn4" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.558437 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8npfx"] Dec 04 13:49:37 crc kubenswrapper[4979]: E1204 13:49:37.559158 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" containerName="ceph-client-openstack-openstack-cell1" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.559180 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" containerName="ceph-client-openstack-openstack-cell1" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.559747 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e47e68-9d7a-4202-97ff-eab1b0e2ccba" containerName="ceph-client-openstack-openstack-cell1" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.560909 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.564073 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.565663 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.566058 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.566228 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.566658 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.571758 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8npfx"] Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.708514 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ssh-key\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.708617 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ceph\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.708639 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-inventory\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.709051 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf790ff5-131c-4db5-b428-2b920b290579-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.709373 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.709477 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slh22\" (UniqueName: \"kubernetes.io/projected/bf790ff5-131c-4db5-b428-2b920b290579-kube-api-access-slh22\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.812479 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ssh-key\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.812596 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ceph\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.812631 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-inventory\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.812790 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf790ff5-131c-4db5-b428-2b920b290579-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.812964 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.813048 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slh22\" (UniqueName: \"kubernetes.io/projected/bf790ff5-131c-4db5-b428-2b920b290579-kube-api-access-slh22\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.813769 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf790ff5-131c-4db5-b428-2b920b290579-ovncontroller-config-0\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.818482 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ssh-key\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.819701 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ovn-combined-ca-bundle\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.819703 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ceph\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.820160 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-inventory\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.832719 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slh22\" (UniqueName: \"kubernetes.io/projected/bf790ff5-131c-4db5-b428-2b920b290579-kube-api-access-slh22\") pod \"ovn-openstack-openstack-cell1-8npfx\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:37 crc kubenswrapper[4979]: I1204 13:49:37.898544 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:49:38 crc kubenswrapper[4979]: I1204 13:49:38.454936 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-openstack-openstack-cell1-8npfx"] Dec 04 13:49:38 crc kubenswrapper[4979]: I1204 13:49:38.493787 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8npfx" event={"ID":"bf790ff5-131c-4db5-b428-2b920b290579","Type":"ContainerStarted","Data":"44c8b5da41dedb649f9bb72edf5f2733c789eb59bdd8435677cade37ed579d45"} Dec 04 13:49:39 crc kubenswrapper[4979]: I1204 13:49:39.505011 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8npfx" event={"ID":"bf790ff5-131c-4db5-b428-2b920b290579","Type":"ContainerStarted","Data":"7aaa251d4a225f4b09f57dbe0967ee68a1815fddb66239f3ab33c7effd4c3a4d"} Dec 04 13:49:39 crc kubenswrapper[4979]: I1204 13:49:39.535901 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-openstack-openstack-cell1-8npfx" podStartSLOduration=1.895829482 podStartE2EDuration="2.535881237s" podCreationTimestamp="2025-12-04 13:49:37 +0000 UTC" firstStartedPulling="2025-12-04 13:49:38.467654559 +0000 UTC m=+7602.741950373" lastFinishedPulling="2025-12-04 13:49:39.107706314 +0000 UTC m=+7603.382002128" observedRunningTime="2025-12-04 13:49:39.528733926 +0000 UTC m=+7603.803029730" watchObservedRunningTime="2025-12-04 13:49:39.535881237 +0000 UTC m=+7603.810177041" Dec 04 13:50:46 crc kubenswrapper[4979]: I1204 13:50:46.232057 4979 generic.go:334] "Generic (PLEG): container finished" podID="bf790ff5-131c-4db5-b428-2b920b290579" containerID="7aaa251d4a225f4b09f57dbe0967ee68a1815fddb66239f3ab33c7effd4c3a4d" exitCode=0 Dec 04 13:50:46 crc kubenswrapper[4979]: I1204 13:50:46.232155 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8npfx" event={"ID":"bf790ff5-131c-4db5-b428-2b920b290579","Type":"ContainerDied","Data":"7aaa251d4a225f4b09f57dbe0967ee68a1815fddb66239f3ab33c7effd4c3a4d"} Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.767803 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.840253 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ovn-combined-ca-bundle\") pod \"bf790ff5-131c-4db5-b428-2b920b290579\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.840336 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf790ff5-131c-4db5-b428-2b920b290579-ovncontroller-config-0\") pod \"bf790ff5-131c-4db5-b428-2b920b290579\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.840413 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slh22\" (UniqueName: \"kubernetes.io/projected/bf790ff5-131c-4db5-b428-2b920b290579-kube-api-access-slh22\") pod \"bf790ff5-131c-4db5-b428-2b920b290579\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.840565 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ssh-key\") pod \"bf790ff5-131c-4db5-b428-2b920b290579\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.841261 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ceph\") pod \"bf790ff5-131c-4db5-b428-2b920b290579\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.841416 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-inventory\") pod \"bf790ff5-131c-4db5-b428-2b920b290579\" (UID: \"bf790ff5-131c-4db5-b428-2b920b290579\") " Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.846721 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ceph" (OuterVolumeSpecName: "ceph") pod "bf790ff5-131c-4db5-b428-2b920b290579" (UID: "bf790ff5-131c-4db5-b428-2b920b290579"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.847057 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf790ff5-131c-4db5-b428-2b920b290579-kube-api-access-slh22" (OuterVolumeSpecName: "kube-api-access-slh22") pod "bf790ff5-131c-4db5-b428-2b920b290579" (UID: "bf790ff5-131c-4db5-b428-2b920b290579"). InnerVolumeSpecName "kube-api-access-slh22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.847226 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "bf790ff5-131c-4db5-b428-2b920b290579" (UID: "bf790ff5-131c-4db5-b428-2b920b290579"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.874156 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf790ff5-131c-4db5-b428-2b920b290579-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "bf790ff5-131c-4db5-b428-2b920b290579" (UID: "bf790ff5-131c-4db5-b428-2b920b290579"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.875911 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-inventory" (OuterVolumeSpecName: "inventory") pod "bf790ff5-131c-4db5-b428-2b920b290579" (UID: "bf790ff5-131c-4db5-b428-2b920b290579"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.888457 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf790ff5-131c-4db5-b428-2b920b290579" (UID: "bf790ff5-131c-4db5-b428-2b920b290579"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.944698 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slh22\" (UniqueName: \"kubernetes.io/projected/bf790ff5-131c-4db5-b428-2b920b290579-kube-api-access-slh22\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.944737 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.944751 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.944765 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.944778 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf790ff5-131c-4db5-b428-2b920b290579-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:47 crc kubenswrapper[4979]: I1204 13:50:47.944788 4979 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/bf790ff5-131c-4db5-b428-2b920b290579-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.254976 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-openstack-openstack-cell1-8npfx" event={"ID":"bf790ff5-131c-4db5-b428-2b920b290579","Type":"ContainerDied","Data":"44c8b5da41dedb649f9bb72edf5f2733c789eb59bdd8435677cade37ed579d45"} Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.255345 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44c8b5da41dedb649f9bb72edf5f2733c789eb59bdd8435677cade37ed579d45" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.255358 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-openstack-openstack-cell1-8npfx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.355223 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-p86wx"] Dec 04 13:50:48 crc kubenswrapper[4979]: E1204 13:50:48.355688 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf790ff5-131c-4db5-b428-2b920b290579" containerName="ovn-openstack-openstack-cell1" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.355709 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf790ff5-131c-4db5-b428-2b920b290579" containerName="ovn-openstack-openstack-cell1" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.355934 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf790ff5-131c-4db5-b428-2b920b290579" containerName="ovn-openstack-openstack-cell1" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.356818 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.363449 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.363697 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.363704 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.364229 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.364358 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.364558 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.370106 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-p86wx"] Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.456262 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.456544 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.457045 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.457398 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.457470 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlz7w\" (UniqueName: \"kubernetes.io/projected/41360a09-d209-4e16-9268-791d9b9afa82-kube-api-access-wlz7w\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.457535 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.457691 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565224 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565473 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565566 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565623 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565647 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlz7w\" (UniqueName: \"kubernetes.io/projected/41360a09-d209-4e16-9268-791d9b9afa82-kube-api-access-wlz7w\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565683 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.565728 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.571594 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.572287 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-nova-metadata-neutron-config-0\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.576880 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ssh-key\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.582804 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ceph\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.584193 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-inventory\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.589044 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.592250 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlz7w\" (UniqueName: \"kubernetes.io/projected/41360a09-d209-4e16-9268-791d9b9afa82-kube-api-access-wlz7w\") pod \"neutron-metadata-openstack-openstack-cell1-p86wx\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:48 crc kubenswrapper[4979]: I1204 13:50:48.677022 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:50:49 crc kubenswrapper[4979]: I1204 13:50:49.217123 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-openstack-openstack-cell1-p86wx"] Dec 04 13:50:49 crc kubenswrapper[4979]: I1204 13:50:49.267419 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" event={"ID":"41360a09-d209-4e16-9268-791d9b9afa82","Type":"ContainerStarted","Data":"6c496829752c3e9f1d1f3e6bdc0462c71927698ebdd3b7c7992aa2eaef5bf697"} Dec 04 13:50:51 crc kubenswrapper[4979]: I1204 13:50:51.289941 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" event={"ID":"41360a09-d209-4e16-9268-791d9b9afa82","Type":"ContainerStarted","Data":"b26c6e12c398dedd6cc15b6470910fcde80225918b7ed946ec15caca4c64195b"} Dec 04 13:50:51 crc kubenswrapper[4979]: I1204 13:50:51.313448 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" podStartSLOduration=2.4885748100000002 podStartE2EDuration="3.313430761s" podCreationTimestamp="2025-12-04 13:50:48 +0000 UTC" firstStartedPulling="2025-12-04 13:50:49.222139926 +0000 UTC m=+7673.496435730" lastFinishedPulling="2025-12-04 13:50:50.046995877 +0000 UTC m=+7674.321291681" observedRunningTime="2025-12-04 13:50:51.307781139 +0000 UTC m=+7675.582076943" watchObservedRunningTime="2025-12-04 13:50:51.313430761 +0000 UTC m=+7675.587726555" Dec 04 13:51:28 crc kubenswrapper[4979]: I1204 13:51:28.041417 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:51:28 crc kubenswrapper[4979]: I1204 13:51:28.042178 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:51:43 crc kubenswrapper[4979]: I1204 13:51:43.823535 4979 generic.go:334] "Generic (PLEG): container finished" podID="41360a09-d209-4e16-9268-791d9b9afa82" containerID="b26c6e12c398dedd6cc15b6470910fcde80225918b7ed946ec15caca4c64195b" exitCode=0 Dec 04 13:51:43 crc kubenswrapper[4979]: I1204 13:51:43.823653 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" event={"ID":"41360a09-d209-4e16-9268-791d9b9afa82","Type":"ContainerDied","Data":"b26c6e12c398dedd6cc15b6470910fcde80225918b7ed946ec15caca4c64195b"} Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.286210 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.455367 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-ovn-metadata-agent-neutron-config-0\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.455510 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-metadata-combined-ca-bundle\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.455676 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-inventory\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.455755 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-nova-metadata-neutron-config-0\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.455977 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ceph\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.456086 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ssh-key\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.456204 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlz7w\" (UniqueName: \"kubernetes.io/projected/41360a09-d209-4e16-9268-791d9b9afa82-kube-api-access-wlz7w\") pod \"41360a09-d209-4e16-9268-791d9b9afa82\" (UID: \"41360a09-d209-4e16-9268-791d9b9afa82\") " Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.461642 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.461732 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41360a09-d209-4e16-9268-791d9b9afa82-kube-api-access-wlz7w" (OuterVolumeSpecName: "kube-api-access-wlz7w") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "kube-api-access-wlz7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.465110 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ceph" (OuterVolumeSpecName: "ceph") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.490553 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-inventory" (OuterVolumeSpecName: "inventory") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.493125 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.493691 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.496289 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "41360a09-d209-4e16-9268-791d9b9afa82" (UID: "41360a09-d209-4e16-9268-791d9b9afa82"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559374 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559414 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559430 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlz7w\" (UniqueName: \"kubernetes.io/projected/41360a09-d209-4e16-9268-791d9b9afa82-kube-api-access-wlz7w\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559444 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559459 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559509 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.559523 4979 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/41360a09-d209-4e16-9268-791d9b9afa82-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.849075 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" event={"ID":"41360a09-d209-4e16-9268-791d9b9afa82","Type":"ContainerDied","Data":"6c496829752c3e9f1d1f3e6bdc0462c71927698ebdd3b7c7992aa2eaef5bf697"} Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.849115 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c496829752c3e9f1d1f3e6bdc0462c71927698ebdd3b7c7992aa2eaef5bf697" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.849177 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-openstack-openstack-cell1-p86wx" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.947915 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-62gcp"] Dec 04 13:51:45 crc kubenswrapper[4979]: E1204 13:51:45.948657 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41360a09-d209-4e16-9268-791d9b9afa82" containerName="neutron-metadata-openstack-openstack-cell1" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.948679 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="41360a09-d209-4e16-9268-791d9b9afa82" containerName="neutron-metadata-openstack-openstack-cell1" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.948939 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="41360a09-d209-4e16-9268-791d9b9afa82" containerName="neutron-metadata-openstack-openstack-cell1" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.949680 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.952949 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.953456 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.953664 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.953812 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.954066 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:51:45 crc kubenswrapper[4979]: I1204 13:51:45.981836 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-62gcp"] Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.069411 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.069928 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.070081 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ceph\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.070136 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-inventory\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.070171 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm489\" (UniqueName: \"kubernetes.io/projected/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-kube-api-access-jm489\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.070312 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ssh-key\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.172607 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.172759 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ceph\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.172787 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-inventory\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.172817 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm489\" (UniqueName: \"kubernetes.io/projected/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-kube-api-access-jm489\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.172857 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ssh-key\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.172929 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.176736 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-inventory\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.176767 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-secret-0\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.176971 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-combined-ca-bundle\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.177607 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ssh-key\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.177816 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ceph\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.190512 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm489\" (UniqueName: \"kubernetes.io/projected/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-kube-api-access-jm489\") pod \"libvirt-openstack-openstack-cell1-62gcp\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.274371 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.802446 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-openstack-openstack-cell1-62gcp"] Dec 04 13:51:46 crc kubenswrapper[4979]: I1204 13:51:46.860290 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" event={"ID":"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8","Type":"ContainerStarted","Data":"c689e80b9e7f97ccbe43ecd6b07821158e5d6c4f645eb66d721c2e90b546aca5"} Dec 04 13:51:49 crc kubenswrapper[4979]: I1204 13:51:49.905940 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" event={"ID":"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8","Type":"ContainerStarted","Data":"af7a7f8fd430096ab620df490c9d9132b7b914a610d3b88d4a3d56636dd3e258"} Dec 04 13:51:49 crc kubenswrapper[4979]: I1204 13:51:49.932185 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" podStartSLOduration=2.780301192 podStartE2EDuration="4.932164923s" podCreationTimestamp="2025-12-04 13:51:45 +0000 UTC" firstStartedPulling="2025-12-04 13:51:46.806143376 +0000 UTC m=+7731.080439200" lastFinishedPulling="2025-12-04 13:51:48.958007127 +0000 UTC m=+7733.232302931" observedRunningTime="2025-12-04 13:51:49.924072135 +0000 UTC m=+7734.198367949" watchObservedRunningTime="2025-12-04 13:51:49.932164923 +0000 UTC m=+7734.206460737" Dec 04 13:51:58 crc kubenswrapper[4979]: I1204 13:51:58.041407 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:51:58 crc kubenswrapper[4979]: I1204 13:51:58.042057 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.040828 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.041551 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.041617 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.042740 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.042826 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" gracePeriod=600 Dec 04 13:52:28 crc kubenswrapper[4979]: E1204 13:52:28.251016 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.357232 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" exitCode=0 Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.357320 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3"} Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.357378 4979 scope.go:117] "RemoveContainer" containerID="4d83d72a3f1cfee91bd625d5ef34b5bb09b87f3b468acc46be5ce89bc808d8ac" Dec 04 13:52:28 crc kubenswrapper[4979]: I1204 13:52:28.358094 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:52:28 crc kubenswrapper[4979]: E1204 13:52:28.358405 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.572593 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v5tgn"] Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.580072 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.593776 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v5tgn"] Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.741182 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-catalog-content\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.741446 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-utilities\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.741506 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scp69\" (UniqueName: \"kubernetes.io/projected/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-kube-api-access-scp69\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.843637 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-utilities\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.843703 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scp69\" (UniqueName: \"kubernetes.io/projected/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-kube-api-access-scp69\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.843760 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-catalog-content\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.844080 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-utilities\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.844255 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-catalog-content\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.863511 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scp69\" (UniqueName: \"kubernetes.io/projected/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-kube-api-access-scp69\") pod \"redhat-operators-v5tgn\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:37 crc kubenswrapper[4979]: I1204 13:52:37.909866 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:38 crc kubenswrapper[4979]: I1204 13:52:38.393571 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v5tgn"] Dec 04 13:52:38 crc kubenswrapper[4979]: I1204 13:52:38.488625 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerStarted","Data":"afe64cf59d5684751475da1ad27d79e2986f6faecf8610b558b05ecf8723a5a9"} Dec 04 13:52:39 crc kubenswrapper[4979]: I1204 13:52:39.199994 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:52:39 crc kubenswrapper[4979]: E1204 13:52:39.200381 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:52:39 crc kubenswrapper[4979]: I1204 13:52:39.504781 4979 generic.go:334] "Generic (PLEG): container finished" podID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerID="759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314" exitCode=0 Dec 04 13:52:39 crc kubenswrapper[4979]: I1204 13:52:39.504925 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerDied","Data":"759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314"} Dec 04 13:52:40 crc kubenswrapper[4979]: I1204 13:52:40.517926 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerStarted","Data":"56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836"} Dec 04 13:52:45 crc kubenswrapper[4979]: I1204 13:52:45.577036 4979 generic.go:334] "Generic (PLEG): container finished" podID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerID="56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836" exitCode=0 Dec 04 13:52:45 crc kubenswrapper[4979]: I1204 13:52:45.577122 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerDied","Data":"56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836"} Dec 04 13:52:47 crc kubenswrapper[4979]: I1204 13:52:47.600841 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerStarted","Data":"0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911"} Dec 04 13:52:47 crc kubenswrapper[4979]: I1204 13:52:47.657552 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v5tgn" podStartSLOduration=3.394823568 podStartE2EDuration="10.657523405s" podCreationTimestamp="2025-12-04 13:52:37 +0000 UTC" firstStartedPulling="2025-12-04 13:52:39.508688413 +0000 UTC m=+7783.782984217" lastFinishedPulling="2025-12-04 13:52:46.77138826 +0000 UTC m=+7791.045684054" observedRunningTime="2025-12-04 13:52:47.624480569 +0000 UTC m=+7791.898776413" watchObservedRunningTime="2025-12-04 13:52:47.657523405 +0000 UTC m=+7791.931819209" Dec 04 13:52:47 crc kubenswrapper[4979]: I1204 13:52:47.910643 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:47 crc kubenswrapper[4979]: I1204 13:52:47.910735 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:48 crc kubenswrapper[4979]: I1204 13:52:48.958361 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v5tgn" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="registry-server" probeResult="failure" output=< Dec 04 13:52:48 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 13:52:48 crc kubenswrapper[4979]: > Dec 04 13:52:50 crc kubenswrapper[4979]: I1204 13:52:50.200387 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:52:50 crc kubenswrapper[4979]: E1204 13:52:50.201654 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:52:57 crc kubenswrapper[4979]: I1204 13:52:57.966004 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:58 crc kubenswrapper[4979]: I1204 13:52:58.032749 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:52:58 crc kubenswrapper[4979]: I1204 13:52:58.212470 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v5tgn"] Dec 04 13:52:59 crc kubenswrapper[4979]: I1204 13:52:59.731046 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v5tgn" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="registry-server" containerID="cri-o://0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911" gracePeriod=2 Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.229981 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.350386 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-utilities\") pod \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.350474 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-catalog-content\") pod \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.350514 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scp69\" (UniqueName: \"kubernetes.io/projected/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-kube-api-access-scp69\") pod \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\" (UID: \"769c5a6e-f79e-4471-a5b6-1e151c9afbc4\") " Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.351523 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-utilities" (OuterVolumeSpecName: "utilities") pod "769c5a6e-f79e-4471-a5b6-1e151c9afbc4" (UID: "769c5a6e-f79e-4471-a5b6-1e151c9afbc4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.356364 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-kube-api-access-scp69" (OuterVolumeSpecName: "kube-api-access-scp69") pod "769c5a6e-f79e-4471-a5b6-1e151c9afbc4" (UID: "769c5a6e-f79e-4471-a5b6-1e151c9afbc4"). InnerVolumeSpecName "kube-api-access-scp69". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.453735 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.453782 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scp69\" (UniqueName: \"kubernetes.io/projected/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-kube-api-access-scp69\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.457173 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "769c5a6e-f79e-4471-a5b6-1e151c9afbc4" (UID: "769c5a6e-f79e-4471-a5b6-1e151c9afbc4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.555983 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/769c5a6e-f79e-4471-a5b6-1e151c9afbc4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.747065 4979 generic.go:334] "Generic (PLEG): container finished" podID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerID="0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911" exitCode=0 Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.747107 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerDied","Data":"0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911"} Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.747137 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v5tgn" event={"ID":"769c5a6e-f79e-4471-a5b6-1e151c9afbc4","Type":"ContainerDied","Data":"afe64cf59d5684751475da1ad27d79e2986f6faecf8610b558b05ecf8723a5a9"} Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.747159 4979 scope.go:117] "RemoveContainer" containerID="0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.747208 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v5tgn" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.773084 4979 scope.go:117] "RemoveContainer" containerID="56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.824466 4979 scope.go:117] "RemoveContainer" containerID="759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.835423 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v5tgn"] Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.856143 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v5tgn"] Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.862142 4979 scope.go:117] "RemoveContainer" containerID="0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911" Dec 04 13:53:00 crc kubenswrapper[4979]: E1204 13:53:00.862698 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911\": container with ID starting with 0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911 not found: ID does not exist" containerID="0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.862737 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911"} err="failed to get container status \"0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911\": rpc error: code = NotFound desc = could not find container \"0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911\": container with ID starting with 0ddb7829e902809d5c5dacdd83a5d2728957cb88f4befb7cfc96ede6ed9cc911 not found: ID does not exist" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.862766 4979 scope.go:117] "RemoveContainer" containerID="56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836" Dec 04 13:53:00 crc kubenswrapper[4979]: E1204 13:53:00.863700 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836\": container with ID starting with 56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836 not found: ID does not exist" containerID="56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.863733 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836"} err="failed to get container status \"56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836\": rpc error: code = NotFound desc = could not find container \"56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836\": container with ID starting with 56f0a29d3519ecb81619ec54fa0f1a6b6c5dd2627b0d7b3e854ced93a21a4836 not found: ID does not exist" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.863752 4979 scope.go:117] "RemoveContainer" containerID="759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314" Dec 04 13:53:00 crc kubenswrapper[4979]: E1204 13:53:00.864139 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314\": container with ID starting with 759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314 not found: ID does not exist" containerID="759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314" Dec 04 13:53:00 crc kubenswrapper[4979]: I1204 13:53:00.864198 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314"} err="failed to get container status \"759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314\": rpc error: code = NotFound desc = could not find container \"759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314\": container with ID starting with 759b1d6dfd15de55079fd07bc70da1a864515f92381d22bd0b39ea8fa7a20314 not found: ID does not exist" Dec 04 13:53:01 crc kubenswrapper[4979]: I1204 13:53:01.199185 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:53:01 crc kubenswrapper[4979]: E1204 13:53:01.199544 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:53:02 crc kubenswrapper[4979]: I1204 13:53:02.210729 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" path="/var/lib/kubelet/pods/769c5a6e-f79e-4471-a5b6-1e151c9afbc4/volumes" Dec 04 13:53:12 crc kubenswrapper[4979]: I1204 13:53:12.200064 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:53:12 crc kubenswrapper[4979]: E1204 13:53:12.200907 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:53:25 crc kubenswrapper[4979]: I1204 13:53:25.199583 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:53:25 crc kubenswrapper[4979]: E1204 13:53:25.200712 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:53:37 crc kubenswrapper[4979]: I1204 13:53:37.199462 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:53:37 crc kubenswrapper[4979]: E1204 13:53:37.200488 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:53:49 crc kubenswrapper[4979]: I1204 13:53:49.199185 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:53:49 crc kubenswrapper[4979]: E1204 13:53:49.200066 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:54:03 crc kubenswrapper[4979]: I1204 13:54:03.199632 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:54:03 crc kubenswrapper[4979]: E1204 13:54:03.200718 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:54:14 crc kubenswrapper[4979]: I1204 13:54:14.199485 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:54:14 crc kubenswrapper[4979]: E1204 13:54:14.200980 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:54:27 crc kubenswrapper[4979]: I1204 13:54:27.200419 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:54:27 crc kubenswrapper[4979]: E1204 13:54:27.201673 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:54:39 crc kubenswrapper[4979]: I1204 13:54:39.199384 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:54:39 crc kubenswrapper[4979]: E1204 13:54:39.200332 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:54:51 crc kubenswrapper[4979]: I1204 13:54:51.199594 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:54:51 crc kubenswrapper[4979]: E1204 13:54:51.200623 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:55:04 crc kubenswrapper[4979]: I1204 13:55:04.200445 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:55:04 crc kubenswrapper[4979]: E1204 13:55:04.201395 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:55:14 crc kubenswrapper[4979]: I1204 13:55:14.579581 4979 scope.go:117] "RemoveContainer" containerID="4ab9cbf787ecbd532648af29f37046c2af78fc369a47e57c0de39ec8ab3c3baf" Dec 04 13:55:14 crc kubenswrapper[4979]: I1204 13:55:14.609770 4979 scope.go:117] "RemoveContainer" containerID="871f573f5bbc1984ccae5ee2eb1a177d1167883007d812493cd40e2cfdcf46d8" Dec 04 13:55:14 crc kubenswrapper[4979]: I1204 13:55:14.632492 4979 scope.go:117] "RemoveContainer" containerID="6f320ae6dcb891181689d6e2eb8b3697c45d9e01f4b7649cb6d45cad4bce566a" Dec 04 13:55:18 crc kubenswrapper[4979]: I1204 13:55:18.199559 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:55:18 crc kubenswrapper[4979]: E1204 13:55:18.200211 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:55:32 crc kubenswrapper[4979]: I1204 13:55:32.200526 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:55:32 crc kubenswrapper[4979]: E1204 13:55:32.201866 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.006854 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4pl5j"] Dec 04 13:55:37 crc kubenswrapper[4979]: E1204 13:55:37.007928 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="extract-content" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.007944 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="extract-content" Dec 04 13:55:37 crc kubenswrapper[4979]: E1204 13:55:37.007983 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="extract-utilities" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.007991 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="extract-utilities" Dec 04 13:55:37 crc kubenswrapper[4979]: E1204 13:55:37.008003 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="registry-server" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.008011 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="registry-server" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.008276 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="769c5a6e-f79e-4471-a5b6-1e151c9afbc4" containerName="registry-server" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.010167 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.023902 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pl5j"] Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.137715 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-utilities\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.137781 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbnfm\" (UniqueName: \"kubernetes.io/projected/f0cf069c-dab1-46fb-aaa3-629a8b340aed-kube-api-access-rbnfm\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.138234 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-catalog-content\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.239849 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-utilities\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.239922 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbnfm\" (UniqueName: \"kubernetes.io/projected/f0cf069c-dab1-46fb-aaa3-629a8b340aed-kube-api-access-rbnfm\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.240043 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-catalog-content\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.240404 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-utilities\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.240505 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-catalog-content\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.260619 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbnfm\" (UniqueName: \"kubernetes.io/projected/f0cf069c-dab1-46fb-aaa3-629a8b340aed-kube-api-access-rbnfm\") pod \"redhat-marketplace-4pl5j\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.344985 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:37 crc kubenswrapper[4979]: I1204 13:55:37.849827 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pl5j"] Dec 04 13:55:37 crc kubenswrapper[4979]: W1204 13:55:37.852913 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0cf069c_dab1_46fb_aaa3_629a8b340aed.slice/crio-cc42750764119f89099a45a92b6ea1ef32f0448595a0ce85e3404cbacf035c65 WatchSource:0}: Error finding container cc42750764119f89099a45a92b6ea1ef32f0448595a0ce85e3404cbacf035c65: Status 404 returned error can't find the container with id cc42750764119f89099a45a92b6ea1ef32f0448595a0ce85e3404cbacf035c65 Dec 04 13:55:38 crc kubenswrapper[4979]: I1204 13:55:38.371228 4979 generic.go:334] "Generic (PLEG): container finished" podID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerID="ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb" exitCode=0 Dec 04 13:55:38 crc kubenswrapper[4979]: I1204 13:55:38.372295 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pl5j" event={"ID":"f0cf069c-dab1-46fb-aaa3-629a8b340aed","Type":"ContainerDied","Data":"ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb"} Dec 04 13:55:38 crc kubenswrapper[4979]: I1204 13:55:38.372426 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pl5j" event={"ID":"f0cf069c-dab1-46fb-aaa3-629a8b340aed","Type":"ContainerStarted","Data":"cc42750764119f89099a45a92b6ea1ef32f0448595a0ce85e3404cbacf035c65"} Dec 04 13:55:38 crc kubenswrapper[4979]: I1204 13:55:38.373931 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 13:55:40 crc kubenswrapper[4979]: I1204 13:55:40.390710 4979 generic.go:334] "Generic (PLEG): container finished" podID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerID="1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760" exitCode=0 Dec 04 13:55:40 crc kubenswrapper[4979]: I1204 13:55:40.390753 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pl5j" event={"ID":"f0cf069c-dab1-46fb-aaa3-629a8b340aed","Type":"ContainerDied","Data":"1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760"} Dec 04 13:55:41 crc kubenswrapper[4979]: I1204 13:55:41.410909 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pl5j" event={"ID":"f0cf069c-dab1-46fb-aaa3-629a8b340aed","Type":"ContainerStarted","Data":"86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd"} Dec 04 13:55:43 crc kubenswrapper[4979]: I1204 13:55:43.199547 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:55:43 crc kubenswrapper[4979]: E1204 13:55:43.200135 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:55:47 crc kubenswrapper[4979]: I1204 13:55:47.345439 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:47 crc kubenswrapper[4979]: I1204 13:55:47.347763 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:47 crc kubenswrapper[4979]: I1204 13:55:47.391804 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:47 crc kubenswrapper[4979]: I1204 13:55:47.415792 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4pl5j" podStartSLOduration=8.952515974 podStartE2EDuration="11.415770317s" podCreationTimestamp="2025-12-04 13:55:36 +0000 UTC" firstStartedPulling="2025-12-04 13:55:38.373701359 +0000 UTC m=+7962.647997163" lastFinishedPulling="2025-12-04 13:55:40.836955702 +0000 UTC m=+7965.111251506" observedRunningTime="2025-12-04 13:55:41.437682042 +0000 UTC m=+7965.711977846" watchObservedRunningTime="2025-12-04 13:55:47.415770317 +0000 UTC m=+7971.690066121" Dec 04 13:55:47 crc kubenswrapper[4979]: I1204 13:55:47.510332 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:47 crc kubenswrapper[4979]: I1204 13:55:47.630457 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pl5j"] Dec 04 13:55:49 crc kubenswrapper[4979]: I1204 13:55:49.478873 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4pl5j" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="registry-server" containerID="cri-o://86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd" gracePeriod=2 Dec 04 13:55:49 crc kubenswrapper[4979]: I1204 13:55:49.983870 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.126319 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-utilities\") pod \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.126868 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-catalog-content\") pod \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.126908 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbnfm\" (UniqueName: \"kubernetes.io/projected/f0cf069c-dab1-46fb-aaa3-629a8b340aed-kube-api-access-rbnfm\") pod \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\" (UID: \"f0cf069c-dab1-46fb-aaa3-629a8b340aed\") " Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.127401 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-utilities" (OuterVolumeSpecName: "utilities") pod "f0cf069c-dab1-46fb-aaa3-629a8b340aed" (UID: "f0cf069c-dab1-46fb-aaa3-629a8b340aed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.128684 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.133530 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0cf069c-dab1-46fb-aaa3-629a8b340aed-kube-api-access-rbnfm" (OuterVolumeSpecName: "kube-api-access-rbnfm") pod "f0cf069c-dab1-46fb-aaa3-629a8b340aed" (UID: "f0cf069c-dab1-46fb-aaa3-629a8b340aed"). InnerVolumeSpecName "kube-api-access-rbnfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.146634 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0cf069c-dab1-46fb-aaa3-629a8b340aed" (UID: "f0cf069c-dab1-46fb-aaa3-629a8b340aed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.230561 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cf069c-dab1-46fb-aaa3-629a8b340aed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.230588 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbnfm\" (UniqueName: \"kubernetes.io/projected/f0cf069c-dab1-46fb-aaa3-629a8b340aed-kube-api-access-rbnfm\") on node \"crc\" DevicePath \"\"" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.490095 4979 generic.go:334] "Generic (PLEG): container finished" podID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerID="86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd" exitCode=0 Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.490146 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pl5j" event={"ID":"f0cf069c-dab1-46fb-aaa3-629a8b340aed","Type":"ContainerDied","Data":"86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd"} Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.490166 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4pl5j" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.490181 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4pl5j" event={"ID":"f0cf069c-dab1-46fb-aaa3-629a8b340aed","Type":"ContainerDied","Data":"cc42750764119f89099a45a92b6ea1ef32f0448595a0ce85e3404cbacf035c65"} Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.490205 4979 scope.go:117] "RemoveContainer" containerID="86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.516460 4979 scope.go:117] "RemoveContainer" containerID="1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.518430 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pl5j"] Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.530537 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4pl5j"] Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.539365 4979 scope.go:117] "RemoveContainer" containerID="ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.590267 4979 scope.go:117] "RemoveContainer" containerID="86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd" Dec 04 13:55:50 crc kubenswrapper[4979]: E1204 13:55:50.590874 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd\": container with ID starting with 86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd not found: ID does not exist" containerID="86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.590920 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd"} err="failed to get container status \"86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd\": rpc error: code = NotFound desc = could not find container \"86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd\": container with ID starting with 86128e8d5f1175e9b3ece8935f8232d5a1dbf8d2f76b35aceab961a52a5246cd not found: ID does not exist" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.590953 4979 scope.go:117] "RemoveContainer" containerID="1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760" Dec 04 13:55:50 crc kubenswrapper[4979]: E1204 13:55:50.591495 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760\": container with ID starting with 1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760 not found: ID does not exist" containerID="1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.591532 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760"} err="failed to get container status \"1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760\": rpc error: code = NotFound desc = could not find container \"1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760\": container with ID starting with 1d7a9dfb143172d86173ed68f6740eebb3e5912fc46efc4bd4cbb27a67d3a760 not found: ID does not exist" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.591559 4979 scope.go:117] "RemoveContainer" containerID="ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb" Dec 04 13:55:50 crc kubenswrapper[4979]: E1204 13:55:50.591869 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb\": container with ID starting with ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb not found: ID does not exist" containerID="ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb" Dec 04 13:55:50 crc kubenswrapper[4979]: I1204 13:55:50.591917 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb"} err="failed to get container status \"ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb\": rpc error: code = NotFound desc = could not find container \"ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb\": container with ID starting with ce3ee9ef3a56290d89e287643e12572836f5f0eeb0455cd5bf0f848bbc0a76fb not found: ID does not exist" Dec 04 13:55:52 crc kubenswrapper[4979]: I1204 13:55:52.219329 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" path="/var/lib/kubelet/pods/f0cf069c-dab1-46fb-aaa3-629a8b340aed/volumes" Dec 04 13:55:57 crc kubenswrapper[4979]: I1204 13:55:57.199196 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:55:57 crc kubenswrapper[4979]: E1204 13:55:57.200046 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:56:09 crc kubenswrapper[4979]: I1204 13:56:09.199392 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:56:09 crc kubenswrapper[4979]: E1204 13:56:09.200292 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:56:22 crc kubenswrapper[4979]: I1204 13:56:22.199871 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:56:22 crc kubenswrapper[4979]: E1204 13:56:22.200810 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:56:22 crc kubenswrapper[4979]: I1204 13:56:22.818599 4979 generic.go:334] "Generic (PLEG): container finished" podID="6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" containerID="af7a7f8fd430096ab620df490c9d9132b7b914a610d3b88d4a3d56636dd3e258" exitCode=0 Dec 04 13:56:22 crc kubenswrapper[4979]: I1204 13:56:22.818834 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" event={"ID":"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8","Type":"ContainerDied","Data":"af7a7f8fd430096ab620df490c9d9132b7b914a610d3b88d4a3d56636dd3e258"} Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.338468 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.512338 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ssh-key\") pod \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.512789 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-combined-ca-bundle\") pod \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.512866 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ceph\") pod \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.513011 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-inventory\") pod \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.513091 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-secret-0\") pod \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.513159 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm489\" (UniqueName: \"kubernetes.io/projected/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-kube-api-access-jm489\") pod \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\" (UID: \"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8\") " Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.532145 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ceph" (OuterVolumeSpecName: "ceph") pod "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" (UID: "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.533501 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" (UID: "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.533777 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-kube-api-access-jm489" (OuterVolumeSpecName: "kube-api-access-jm489") pod "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" (UID: "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8"). InnerVolumeSpecName "kube-api-access-jm489". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.550580 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" (UID: "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.558902 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-inventory" (OuterVolumeSpecName: "inventory") pod "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" (UID: "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.559207 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" (UID: "6e6e1861-f679-4685-9e3a-9bf6b9fb00a8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.616816 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.616861 4979 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.616880 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm489\" (UniqueName: \"kubernetes.io/projected/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-kube-api-access-jm489\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.616896 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.616909 4979 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.616921 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6e6e1861-f679-4685-9e3a-9bf6b9fb00a8-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.846556 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" event={"ID":"6e6e1861-f679-4685-9e3a-9bf6b9fb00a8","Type":"ContainerDied","Data":"c689e80b9e7f97ccbe43ecd6b07821158e5d6c4f645eb66d721c2e90b546aca5"} Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.846613 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c689e80b9e7f97ccbe43ecd6b07821158e5d6c4f645eb66d721c2e90b546aca5" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.846718 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-openstack-openstack-cell1-62gcp" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.935218 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8cfz2"] Dec 04 13:56:24 crc kubenswrapper[4979]: E1204 13:56:24.935738 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="extract-content" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.935752 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="extract-content" Dec 04 13:56:24 crc kubenswrapper[4979]: E1204 13:56:24.935775 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" containerName="libvirt-openstack-openstack-cell1" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.935781 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" containerName="libvirt-openstack-openstack-cell1" Dec 04 13:56:24 crc kubenswrapper[4979]: E1204 13:56:24.935799 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="extract-utilities" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.935806 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="extract-utilities" Dec 04 13:56:24 crc kubenswrapper[4979]: E1204 13:56:24.935819 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="registry-server" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.935826 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="registry-server" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.936036 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e6e1861-f679-4685-9e3a-9bf6b9fb00a8" containerName="libvirt-openstack-openstack-cell1" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.936061 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0cf069c-dab1-46fb-aaa3-629a8b340aed" containerName="registry-server" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.937036 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.939375 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.939558 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.939830 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.940916 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.941021 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.941115 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.941024 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 13:56:24 crc kubenswrapper[4979]: I1204 13:56:24.947803 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8cfz2"] Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024470 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024553 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024587 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024611 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024708 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ceph\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024819 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024853 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024873 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.024936 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.025056 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgw25\" (UniqueName: \"kubernetes.io/projected/3250581b-c62c-4c80-9738-c997e7600719-kube-api-access-bgw25\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.025147 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127076 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127147 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127170 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127216 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127263 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgw25\" (UniqueName: \"kubernetes.io/projected/3250581b-c62c-4c80-9738-c997e7600719-kube-api-access-bgw25\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127347 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127433 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127456 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127480 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127507 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.127568 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ceph\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.129489 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.129516 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.131363 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ceph\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.131726 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-inventory\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.132240 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ssh-key\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.133669 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.133898 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.135966 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.136233 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.136557 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.147429 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgw25\" (UniqueName: \"kubernetes.io/projected/3250581b-c62c-4c80-9738-c997e7600719-kube-api-access-bgw25\") pod \"nova-cell1-openstack-openstack-cell1-8cfz2\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.266703 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.794388 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-openstack-cell1-8cfz2"] Dec 04 13:56:25 crc kubenswrapper[4979]: I1204 13:56:25.861963 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" event={"ID":"3250581b-c62c-4c80-9738-c997e7600719","Type":"ContainerStarted","Data":"eb85cbe2e0db065b6e57a72a84f2af6b8a3ede86346948940479e8b55168749e"} Dec 04 13:56:26 crc kubenswrapper[4979]: I1204 13:56:26.872616 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" event={"ID":"3250581b-c62c-4c80-9738-c997e7600719","Type":"ContainerStarted","Data":"bd8868ef562896093b9912f1d9b3ffb6254ed280a49c0d7f7309b2483c571da5"} Dec 04 13:56:26 crc kubenswrapper[4979]: I1204 13:56:26.903455 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" podStartSLOduration=2.508063614 podStartE2EDuration="2.903436737s" podCreationTimestamp="2025-12-04 13:56:24 +0000 UTC" firstStartedPulling="2025-12-04 13:56:25.791682371 +0000 UTC m=+8010.065978175" lastFinishedPulling="2025-12-04 13:56:26.187055504 +0000 UTC m=+8010.461351298" observedRunningTime="2025-12-04 13:56:26.894079916 +0000 UTC m=+8011.168375720" watchObservedRunningTime="2025-12-04 13:56:26.903436737 +0000 UTC m=+8011.177732541" Dec 04 13:56:36 crc kubenswrapper[4979]: I1204 13:56:36.205806 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:56:36 crc kubenswrapper[4979]: E1204 13:56:36.206704 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:56:47 crc kubenswrapper[4979]: I1204 13:56:47.200956 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:56:47 crc kubenswrapper[4979]: E1204 13:56:47.202370 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:56:58 crc kubenswrapper[4979]: I1204 13:56:58.199436 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:56:58 crc kubenswrapper[4979]: E1204 13:56:58.200223 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:57:13 crc kubenswrapper[4979]: I1204 13:57:13.199608 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:57:13 crc kubenswrapper[4979]: E1204 13:57:13.200610 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:57:26 crc kubenswrapper[4979]: I1204 13:57:26.207518 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:57:26 crc kubenswrapper[4979]: E1204 13:57:26.208208 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 13:57:37 crc kubenswrapper[4979]: I1204 13:57:37.199347 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 13:57:38 crc kubenswrapper[4979]: I1204 13:57:38.631083 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"33c6921fae4bfde0599067dbe7d0511a70dccbc78375ca402942f8ac5afb854b"} Dec 04 13:59:00 crc kubenswrapper[4979]: I1204 13:59:00.940507 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g8mhf"] Dec 04 13:59:00 crc kubenswrapper[4979]: I1204 13:59:00.943479 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:00 crc kubenswrapper[4979]: I1204 13:59:00.975105 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g8mhf"] Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.034986 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-utilities\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.035203 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-catalog-content\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.035543 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bm84\" (UniqueName: \"kubernetes.io/projected/9103e0ce-a1c7-486e-9e68-2cf4da85da88-kube-api-access-6bm84\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.137669 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-utilities\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.137741 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-catalog-content\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.137803 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bm84\" (UniqueName: \"kubernetes.io/projected/9103e0ce-a1c7-486e-9e68-2cf4da85da88-kube-api-access-6bm84\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.138477 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-utilities\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.138903 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-catalog-content\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.168768 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bm84\" (UniqueName: \"kubernetes.io/projected/9103e0ce-a1c7-486e-9e68-2cf4da85da88-kube-api-access-6bm84\") pod \"certified-operators-g8mhf\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.263599 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:01 crc kubenswrapper[4979]: I1204 13:59:01.861366 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g8mhf"] Dec 04 13:59:02 crc kubenswrapper[4979]: I1204 13:59:02.459825 4979 generic.go:334] "Generic (PLEG): container finished" podID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerID="846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4" exitCode=0 Dec 04 13:59:02 crc kubenswrapper[4979]: I1204 13:59:02.459905 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerDied","Data":"846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4"} Dec 04 13:59:02 crc kubenswrapper[4979]: I1204 13:59:02.460240 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerStarted","Data":"c599eda568bd50842925cff4438bcc5ba953d145928890ad1e48a1f0cd1a7d57"} Dec 04 13:59:03 crc kubenswrapper[4979]: I1204 13:59:03.473539 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerStarted","Data":"32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a"} Dec 04 13:59:04 crc kubenswrapper[4979]: I1204 13:59:04.486887 4979 generic.go:334] "Generic (PLEG): container finished" podID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerID="32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a" exitCode=0 Dec 04 13:59:04 crc kubenswrapper[4979]: I1204 13:59:04.486955 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerDied","Data":"32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a"} Dec 04 13:59:05 crc kubenswrapper[4979]: I1204 13:59:05.501667 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerStarted","Data":"0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a"} Dec 04 13:59:05 crc kubenswrapper[4979]: I1204 13:59:05.527764 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-g8mhf" podStartSLOduration=2.7958453260000002 podStartE2EDuration="5.527738362s" podCreationTimestamp="2025-12-04 13:59:00 +0000 UTC" firstStartedPulling="2025-12-04 13:59:02.46325478 +0000 UTC m=+8166.737550594" lastFinishedPulling="2025-12-04 13:59:05.195147826 +0000 UTC m=+8169.469443630" observedRunningTime="2025-12-04 13:59:05.519047504 +0000 UTC m=+8169.793343318" watchObservedRunningTime="2025-12-04 13:59:05.527738362 +0000 UTC m=+8169.802034166" Dec 04 13:59:11 crc kubenswrapper[4979]: I1204 13:59:11.263821 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:11 crc kubenswrapper[4979]: I1204 13:59:11.264449 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:11 crc kubenswrapper[4979]: I1204 13:59:11.311104 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:11 crc kubenswrapper[4979]: I1204 13:59:11.620384 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:11 crc kubenswrapper[4979]: I1204 13:59:11.667481 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g8mhf"] Dec 04 13:59:13 crc kubenswrapper[4979]: I1204 13:59:13.592811 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-g8mhf" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="registry-server" containerID="cri-o://0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a" gracePeriod=2 Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.081564 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.139616 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bm84\" (UniqueName: \"kubernetes.io/projected/9103e0ce-a1c7-486e-9e68-2cf4da85da88-kube-api-access-6bm84\") pod \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.139801 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-utilities\") pod \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.139847 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-catalog-content\") pod \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\" (UID: \"9103e0ce-a1c7-486e-9e68-2cf4da85da88\") " Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.140764 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-utilities" (OuterVolumeSpecName: "utilities") pod "9103e0ce-a1c7-486e-9e68-2cf4da85da88" (UID: "9103e0ce-a1c7-486e-9e68-2cf4da85da88"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.145743 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9103e0ce-a1c7-486e-9e68-2cf4da85da88-kube-api-access-6bm84" (OuterVolumeSpecName: "kube-api-access-6bm84") pod "9103e0ce-a1c7-486e-9e68-2cf4da85da88" (UID: "9103e0ce-a1c7-486e-9e68-2cf4da85da88"). InnerVolumeSpecName "kube-api-access-6bm84". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.202101 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9103e0ce-a1c7-486e-9e68-2cf4da85da88" (UID: "9103e0ce-a1c7-486e-9e68-2cf4da85da88"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.244265 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.244308 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9103e0ce-a1c7-486e-9e68-2cf4da85da88-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.244322 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bm84\" (UniqueName: \"kubernetes.io/projected/9103e0ce-a1c7-486e-9e68-2cf4da85da88-kube-api-access-6bm84\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.604799 4979 generic.go:334] "Generic (PLEG): container finished" podID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerID="0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a" exitCode=0 Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.604831 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g8mhf" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.604862 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerDied","Data":"0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a"} Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.605213 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-g8mhf" event={"ID":"9103e0ce-a1c7-486e-9e68-2cf4da85da88","Type":"ContainerDied","Data":"c599eda568bd50842925cff4438bcc5ba953d145928890ad1e48a1f0cd1a7d57"} Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.605231 4979 scope.go:117] "RemoveContainer" containerID="0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.628405 4979 scope.go:117] "RemoveContainer" containerID="32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.633242 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-g8mhf"] Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.647209 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-g8mhf"] Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.652202 4979 scope.go:117] "RemoveContainer" containerID="846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.695256 4979 scope.go:117] "RemoveContainer" containerID="0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a" Dec 04 13:59:14 crc kubenswrapper[4979]: E1204 13:59:14.695745 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a\": container with ID starting with 0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a not found: ID does not exist" containerID="0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.695779 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a"} err="failed to get container status \"0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a\": rpc error: code = NotFound desc = could not find container \"0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a\": container with ID starting with 0cde329ea750a816e17d97ca9138b52c307c34ae8f3800654586032dc9f2109a not found: ID does not exist" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.695803 4979 scope.go:117] "RemoveContainer" containerID="32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a" Dec 04 13:59:14 crc kubenswrapper[4979]: E1204 13:59:14.696622 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a\": container with ID starting with 32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a not found: ID does not exist" containerID="32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.696646 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a"} err="failed to get container status \"32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a\": rpc error: code = NotFound desc = could not find container \"32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a\": container with ID starting with 32e3c1ac1812313056654bbece98e34be8a3d2a9ced60a3c5019bdc0dd33003a not found: ID does not exist" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.696660 4979 scope.go:117] "RemoveContainer" containerID="846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4" Dec 04 13:59:14 crc kubenswrapper[4979]: E1204 13:59:14.696986 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4\": container with ID starting with 846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4 not found: ID does not exist" containerID="846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4" Dec 04 13:59:14 crc kubenswrapper[4979]: I1204 13:59:14.697030 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4"} err="failed to get container status \"846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4\": rpc error: code = NotFound desc = could not find container \"846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4\": container with ID starting with 846c5204890c02824546749786cde3f6c60bfef1d1d418e002028e29f74e93f4 not found: ID does not exist" Dec 04 13:59:16 crc kubenswrapper[4979]: I1204 13:59:16.211055 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" path="/var/lib/kubelet/pods/9103e0ce-a1c7-486e-9e68-2cf4da85da88/volumes" Dec 04 13:59:36 crc kubenswrapper[4979]: I1204 13:59:36.834648 4979 generic.go:334] "Generic (PLEG): container finished" podID="3250581b-c62c-4c80-9738-c997e7600719" containerID="bd8868ef562896093b9912f1d9b3ffb6254ed280a49c0d7f7309b2483c571da5" exitCode=0 Dec 04 13:59:36 crc kubenswrapper[4979]: I1204 13:59:36.834710 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" event={"ID":"3250581b-c62c-4c80-9738-c997e7600719","Type":"ContainerDied","Data":"bd8868ef562896093b9912f1d9b3ffb6254ed280a49c0d7f7309b2483c571da5"} Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.326625 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.402330 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-1\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.402460 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-0\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.402584 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-1\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.402648 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-combined-ca-bundle\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.402722 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-0\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.404354 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-inventory\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.404431 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ceph\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.404474 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-0\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.404635 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgw25\" (UniqueName: \"kubernetes.io/projected/3250581b-c62c-4c80-9738-c997e7600719-kube-api-access-bgw25\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.404743 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ssh-key\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.404780 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-1\") pod \"3250581b-c62c-4c80-9738-c997e7600719\" (UID: \"3250581b-c62c-4c80-9738-c997e7600719\") " Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.414807 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.415549 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3250581b-c62c-4c80-9738-c997e7600719-kube-api-access-bgw25" (OuterVolumeSpecName: "kube-api-access-bgw25") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "kube-api-access-bgw25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.418710 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ceph" (OuterVolumeSpecName: "ceph") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.435759 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.436514 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.438466 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.438646 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.447082 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.459879 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.460523 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-inventory" (OuterVolumeSpecName: "inventory") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.461859 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "3250581b-c62c-4c80-9738-c997e7600719" (UID: "3250581b-c62c-4c80-9738-c997e7600719"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516730 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgw25\" (UniqueName: \"kubernetes.io/projected/3250581b-c62c-4c80-9738-c997e7600719-kube-api-access-bgw25\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516771 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516791 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516805 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516819 4979 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516831 4979 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516848 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516861 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/3250581b-c62c-4c80-9738-c997e7600719-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516873 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516884 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.516899 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/3250581b-c62c-4c80-9738-c997e7600719-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.859728 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" event={"ID":"3250581b-c62c-4c80-9738-c997e7600719","Type":"ContainerDied","Data":"eb85cbe2e0db065b6e57a72a84f2af6b8a3ede86346948940479e8b55168749e"} Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.860134 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb85cbe2e0db065b6e57a72a84f2af6b8a3ede86346948940479e8b55168749e" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.859821 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-openstack-cell1-8cfz2" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.962358 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-szvj8"] Dec 04 13:59:38 crc kubenswrapper[4979]: E1204 13:59:38.962954 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="extract-content" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.962976 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="extract-content" Dec 04 13:59:38 crc kubenswrapper[4979]: E1204 13:59:38.963011 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="extract-utilities" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.963020 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="extract-utilities" Dec 04 13:59:38 crc kubenswrapper[4979]: E1204 13:59:38.963034 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3250581b-c62c-4c80-9738-c997e7600719" containerName="nova-cell1-openstack-openstack-cell1" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.963043 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3250581b-c62c-4c80-9738-c997e7600719" containerName="nova-cell1-openstack-openstack-cell1" Dec 04 13:59:38 crc kubenswrapper[4979]: E1204 13:59:38.963062 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="registry-server" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.963069 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="registry-server" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.963327 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3250581b-c62c-4c80-9738-c997e7600719" containerName="nova-cell1-openstack-openstack-cell1" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.963366 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="9103e0ce-a1c7-486e-9e68-2cf4da85da88" containerName="registry-server" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.964444 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.967154 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.967352 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.967364 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.967394 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 13:59:38 crc kubenswrapper[4979]: I1204 13:59:38.968647 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.008234 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-szvj8"] Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031002 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceph\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031053 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ssh-key\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031073 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-inventory\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031319 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031449 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031492 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcr8x\" (UniqueName: \"kubernetes.io/projected/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-kube-api-access-hcr8x\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031692 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.031879 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.133835 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceph\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.133903 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ssh-key\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.133950 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-inventory\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.134604 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.134696 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.134729 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcr8x\" (UniqueName: \"kubernetes.io/projected/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-kube-api-access-hcr8x\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.136432 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.136705 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.138662 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-1\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.138912 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-2\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.139232 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-inventory\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.139288 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ssh-key\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.140063 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-telemetry-combined-ca-bundle\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.140519 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceph\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.140587 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-0\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.159149 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcr8x\" (UniqueName: \"kubernetes.io/projected/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-kube-api-access-hcr8x\") pod \"telemetry-openstack-openstack-cell1-szvj8\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.300057 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.833427 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-openstack-openstack-cell1-szvj8"] Dec 04 13:59:39 crc kubenswrapper[4979]: I1204 13:59:39.871400 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" event={"ID":"3aacedd4-db12-48a8-9ad0-1c95d92b24f4","Type":"ContainerStarted","Data":"8b448fa45abc38d334a7a0a2ffcf01c1c6d11642339c118d00ebd901f1cc1ebd"} Dec 04 13:59:40 crc kubenswrapper[4979]: I1204 13:59:40.893101 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" event={"ID":"3aacedd4-db12-48a8-9ad0-1c95d92b24f4","Type":"ContainerStarted","Data":"820c6a19defb22b1d5cdb8cfd84dc4b85272c1198dfff58a7d4204a85e7fdbfd"} Dec 04 13:59:40 crc kubenswrapper[4979]: I1204 13:59:40.923784 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" podStartSLOduration=2.469318346 podStartE2EDuration="2.923748794s" podCreationTimestamp="2025-12-04 13:59:38 +0000 UTC" firstStartedPulling="2025-12-04 13:59:39.842552441 +0000 UTC m=+8204.116848245" lastFinishedPulling="2025-12-04 13:59:40.296982889 +0000 UTC m=+8204.571278693" observedRunningTime="2025-12-04 13:59:40.912470077 +0000 UTC m=+8205.186765881" watchObservedRunningTime="2025-12-04 13:59:40.923748794 +0000 UTC m=+8205.198044598" Dec 04 13:59:56 crc kubenswrapper[4979]: I1204 13:59:56.856454 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zz6l6"] Dec 04 13:59:56 crc kubenswrapper[4979]: I1204 13:59:56.860240 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:56 crc kubenswrapper[4979]: I1204 13:59:56.865361 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zz6l6"] Dec 04 13:59:56 crc kubenswrapper[4979]: I1204 13:59:56.965327 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-catalog-content\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:56 crc kubenswrapper[4979]: I1204 13:59:56.965393 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spmh4\" (UniqueName: \"kubernetes.io/projected/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-kube-api-access-spmh4\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:56 crc kubenswrapper[4979]: I1204 13:59:56.965624 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-utilities\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.068151 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-utilities\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.068543 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-catalog-content\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.068684 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spmh4\" (UniqueName: \"kubernetes.io/projected/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-kube-api-access-spmh4\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.068907 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-utilities\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.069103 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-catalog-content\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.093452 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spmh4\" (UniqueName: \"kubernetes.io/projected/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-kube-api-access-spmh4\") pod \"community-operators-zz6l6\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.214742 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 13:59:57 crc kubenswrapper[4979]: I1204 13:59:57.819125 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zz6l6"] Dec 04 13:59:58 crc kubenswrapper[4979]: I1204 13:59:58.041230 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 13:59:58 crc kubenswrapper[4979]: I1204 13:59:58.041650 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 13:59:58 crc kubenswrapper[4979]: I1204 13:59:58.098419 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerStarted","Data":"a0d52ea37844f07d81f9759307be835c83a9a02a29fc881092fe793f523976c2"} Dec 04 13:59:59 crc kubenswrapper[4979]: I1204 13:59:59.108636 4979 generic.go:334] "Generic (PLEG): container finished" podID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerID="3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8" exitCode=0 Dec 04 13:59:59 crc kubenswrapper[4979]: I1204 13:59:59.108675 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerDied","Data":"3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8"} Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.156465 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7"] Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.158521 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.160264 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.160399 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.171117 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7"] Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.248737 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c4vj\" (UniqueName: \"kubernetes.io/projected/3af28718-8edc-4ff4-a732-8443f16005ac-kube-api-access-5c4vj\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.248789 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3af28718-8edc-4ff4-a732-8443f16005ac-secret-volume\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.248829 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3af28718-8edc-4ff4-a732-8443f16005ac-config-volume\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.350412 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c4vj\" (UniqueName: \"kubernetes.io/projected/3af28718-8edc-4ff4-a732-8443f16005ac-kube-api-access-5c4vj\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.350477 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3af28718-8edc-4ff4-a732-8443f16005ac-secret-volume\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.350524 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3af28718-8edc-4ff4-a732-8443f16005ac-config-volume\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.352092 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3af28718-8edc-4ff4-a732-8443f16005ac-config-volume\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.368353 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3af28718-8edc-4ff4-a732-8443f16005ac-secret-volume\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.374232 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c4vj\" (UniqueName: \"kubernetes.io/projected/3af28718-8edc-4ff4-a732-8443f16005ac-kube-api-access-5c4vj\") pod \"collect-profiles-29414280-q6jq7\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:00 crc kubenswrapper[4979]: I1204 14:00:00.518762 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:01 crc kubenswrapper[4979]: I1204 14:00:01.009327 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7"] Dec 04 14:00:01 crc kubenswrapper[4979]: I1204 14:00:01.129231 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" event={"ID":"3af28718-8edc-4ff4-a732-8443f16005ac","Type":"ContainerStarted","Data":"d7bd8012540316c0bdee174d2b544f82b3486b0b36eb922426ec9b13ecbbb5c9"} Dec 04 14:00:01 crc kubenswrapper[4979]: I1204 14:00:01.131420 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerStarted","Data":"67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0"} Dec 04 14:00:02 crc kubenswrapper[4979]: I1204 14:00:02.142320 4979 generic.go:334] "Generic (PLEG): container finished" podID="3af28718-8edc-4ff4-a732-8443f16005ac" containerID="74b255918f4c3dfae3ed8b4aea41641d94cfb0da21bab1236c22c7287966bee0" exitCode=0 Dec 04 14:00:02 crc kubenswrapper[4979]: I1204 14:00:02.142384 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" event={"ID":"3af28718-8edc-4ff4-a732-8443f16005ac","Type":"ContainerDied","Data":"74b255918f4c3dfae3ed8b4aea41641d94cfb0da21bab1236c22c7287966bee0"} Dec 04 14:00:02 crc kubenswrapper[4979]: I1204 14:00:02.144476 4979 generic.go:334] "Generic (PLEG): container finished" podID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerID="67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0" exitCode=0 Dec 04 14:00:02 crc kubenswrapper[4979]: I1204 14:00:02.144514 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerDied","Data":"67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0"} Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.514568 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.627523 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c4vj\" (UniqueName: \"kubernetes.io/projected/3af28718-8edc-4ff4-a732-8443f16005ac-kube-api-access-5c4vj\") pod \"3af28718-8edc-4ff4-a732-8443f16005ac\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.627765 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3af28718-8edc-4ff4-a732-8443f16005ac-config-volume\") pod \"3af28718-8edc-4ff4-a732-8443f16005ac\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.627857 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3af28718-8edc-4ff4-a732-8443f16005ac-secret-volume\") pod \"3af28718-8edc-4ff4-a732-8443f16005ac\" (UID: \"3af28718-8edc-4ff4-a732-8443f16005ac\") " Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.628399 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3af28718-8edc-4ff4-a732-8443f16005ac-config-volume" (OuterVolumeSpecName: "config-volume") pod "3af28718-8edc-4ff4-a732-8443f16005ac" (UID: "3af28718-8edc-4ff4-a732-8443f16005ac"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.634147 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af28718-8edc-4ff4-a732-8443f16005ac-kube-api-access-5c4vj" (OuterVolumeSpecName: "kube-api-access-5c4vj") pod "3af28718-8edc-4ff4-a732-8443f16005ac" (UID: "3af28718-8edc-4ff4-a732-8443f16005ac"). InnerVolumeSpecName "kube-api-access-5c4vj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.634631 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3af28718-8edc-4ff4-a732-8443f16005ac-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3af28718-8edc-4ff4-a732-8443f16005ac" (UID: "3af28718-8edc-4ff4-a732-8443f16005ac"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.731189 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3af28718-8edc-4ff4-a732-8443f16005ac-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.731255 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3af28718-8edc-4ff4-a732-8443f16005ac-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:03 crc kubenswrapper[4979]: I1204 14:00:03.731276 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c4vj\" (UniqueName: \"kubernetes.io/projected/3af28718-8edc-4ff4-a732-8443f16005ac-kube-api-access-5c4vj\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.169040 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.169039 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414280-q6jq7" event={"ID":"3af28718-8edc-4ff4-a732-8443f16005ac","Type":"ContainerDied","Data":"d7bd8012540316c0bdee174d2b544f82b3486b0b36eb922426ec9b13ecbbb5c9"} Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.170420 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7bd8012540316c0bdee174d2b544f82b3486b0b36eb922426ec9b13ecbbb5c9" Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.174265 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerStarted","Data":"0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3"} Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.197805 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zz6l6" podStartSLOduration=4.290018708 podStartE2EDuration="8.197789352s" podCreationTimestamp="2025-12-04 13:59:56 +0000 UTC" firstStartedPulling="2025-12-04 13:59:59.111367616 +0000 UTC m=+8223.385663420" lastFinishedPulling="2025-12-04 14:00:03.01913826 +0000 UTC m=+8227.293434064" observedRunningTime="2025-12-04 14:00:04.193194952 +0000 UTC m=+8228.467490806" watchObservedRunningTime="2025-12-04 14:00:04.197789352 +0000 UTC m=+8228.472085156" Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.592761 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw"] Dec 04 14:00:04 crc kubenswrapper[4979]: I1204 14:00:04.606816 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414235-782dw"] Dec 04 14:00:06 crc kubenswrapper[4979]: I1204 14:00:06.218819 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2705356-a364-400a-95e3-8649b6a3e9aa" path="/var/lib/kubelet/pods/b2705356-a364-400a-95e3-8649b6a3e9aa/volumes" Dec 04 14:00:07 crc kubenswrapper[4979]: I1204 14:00:07.215752 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 14:00:07 crc kubenswrapper[4979]: I1204 14:00:07.216327 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 14:00:07 crc kubenswrapper[4979]: I1204 14:00:07.272363 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 14:00:08 crc kubenswrapper[4979]: I1204 14:00:08.266767 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 14:00:08 crc kubenswrapper[4979]: I1204 14:00:08.316991 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zz6l6"] Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.227761 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zz6l6" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="registry-server" containerID="cri-o://0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3" gracePeriod=2 Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.729718 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.900539 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-catalog-content\") pod \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.901010 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-utilities\") pod \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.901117 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spmh4\" (UniqueName: \"kubernetes.io/projected/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-kube-api-access-spmh4\") pod \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\" (UID: \"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761\") " Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.902212 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-utilities" (OuterVolumeSpecName: "utilities") pod "2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" (UID: "2a1743fa-87f6-46d8-ad30-fc5b9ea2e761"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.908540 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-kube-api-access-spmh4" (OuterVolumeSpecName: "kube-api-access-spmh4") pod "2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" (UID: "2a1743fa-87f6-46d8-ad30-fc5b9ea2e761"). InnerVolumeSpecName "kube-api-access-spmh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:00:10 crc kubenswrapper[4979]: I1204 14:00:10.947952 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" (UID: "2a1743fa-87f6-46d8-ad30-fc5b9ea2e761"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.003287 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.003360 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spmh4\" (UniqueName: \"kubernetes.io/projected/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-kube-api-access-spmh4\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.003378 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.241456 4979 generic.go:334] "Generic (PLEG): container finished" podID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerID="0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3" exitCode=0 Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.241516 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerDied","Data":"0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3"} Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.241579 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zz6l6" event={"ID":"2a1743fa-87f6-46d8-ad30-fc5b9ea2e761","Type":"ContainerDied","Data":"a0d52ea37844f07d81f9759307be835c83a9a02a29fc881092fe793f523976c2"} Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.241598 4979 scope.go:117] "RemoveContainer" containerID="0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.242537 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zz6l6" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.271165 4979 scope.go:117] "RemoveContainer" containerID="67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.294642 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zz6l6"] Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.302852 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zz6l6"] Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.325390 4979 scope.go:117] "RemoveContainer" containerID="3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.371405 4979 scope.go:117] "RemoveContainer" containerID="0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3" Dec 04 14:00:11 crc kubenswrapper[4979]: E1204 14:00:11.371966 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3\": container with ID starting with 0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3 not found: ID does not exist" containerID="0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.372019 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3"} err="failed to get container status \"0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3\": rpc error: code = NotFound desc = could not find container \"0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3\": container with ID starting with 0ae0709c9662a7965b6e11567b2e3ecf106b0d9b1a2a2142bfe4a58e175c79b3 not found: ID does not exist" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.372051 4979 scope.go:117] "RemoveContainer" containerID="67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0" Dec 04 14:00:11 crc kubenswrapper[4979]: E1204 14:00:11.372517 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0\": container with ID starting with 67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0 not found: ID does not exist" containerID="67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.372541 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0"} err="failed to get container status \"67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0\": rpc error: code = NotFound desc = could not find container \"67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0\": container with ID starting with 67914efcc73d01ebfb508beb5725632651b6a4d0179cd85a5fbe9b1ea426eba0 not found: ID does not exist" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.372557 4979 scope.go:117] "RemoveContainer" containerID="3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8" Dec 04 14:00:11 crc kubenswrapper[4979]: E1204 14:00:11.372802 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8\": container with ID starting with 3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8 not found: ID does not exist" containerID="3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8" Dec 04 14:00:11 crc kubenswrapper[4979]: I1204 14:00:11.372830 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8"} err="failed to get container status \"3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8\": rpc error: code = NotFound desc = could not find container \"3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8\": container with ID starting with 3db8511d943e4394468f8b64127e59341d2b9aaa30deb8e26b71847e8f6074e8 not found: ID does not exist" Dec 04 14:00:12 crc kubenswrapper[4979]: I1204 14:00:12.210588 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" path="/var/lib/kubelet/pods/2a1743fa-87f6-46d8-ad30-fc5b9ea2e761/volumes" Dec 04 14:00:14 crc kubenswrapper[4979]: I1204 14:00:14.850757 4979 scope.go:117] "RemoveContainer" containerID="6fb6c72a6ff784c768c49fa34ee00ca6c2acfe2df20d84c7c6cc6833efd83b03" Dec 04 14:00:28 crc kubenswrapper[4979]: I1204 14:00:28.041289 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:00:28 crc kubenswrapper[4979]: I1204 14:00:28.042085 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.040950 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.041502 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.041555 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.042407 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"33c6921fae4bfde0599067dbe7d0511a70dccbc78375ca402942f8ac5afb854b"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.042455 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://33c6921fae4bfde0599067dbe7d0511a70dccbc78375ca402942f8ac5afb854b" gracePeriod=600 Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.736844 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="33c6921fae4bfde0599067dbe7d0511a70dccbc78375ca402942f8ac5afb854b" exitCode=0 Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.736910 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"33c6921fae4bfde0599067dbe7d0511a70dccbc78375ca402942f8ac5afb854b"} Dec 04 14:00:58 crc kubenswrapper[4979]: I1204 14:00:58.737251 4979 scope.go:117] "RemoveContainer" containerID="c3942fc6d4db0669f9cca0cc5af29e062680a552b5e4537e774087fa74dcb0e3" Dec 04 14:00:59 crc kubenswrapper[4979]: I1204 14:00:59.750885 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982"} Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.157611 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29414281-drh6b"] Dec 04 14:01:00 crc kubenswrapper[4979]: E1204 14:01:00.158068 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af28718-8edc-4ff4-a732-8443f16005ac" containerName="collect-profiles" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.158085 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af28718-8edc-4ff4-a732-8443f16005ac" containerName="collect-profiles" Dec 04 14:01:00 crc kubenswrapper[4979]: E1204 14:01:00.158109 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="extract-content" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.158115 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="extract-content" Dec 04 14:01:00 crc kubenswrapper[4979]: E1204 14:01:00.158127 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="registry-server" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.158133 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="registry-server" Dec 04 14:01:00 crc kubenswrapper[4979]: E1204 14:01:00.158147 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="extract-utilities" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.158154 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="extract-utilities" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.158360 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af28718-8edc-4ff4-a732-8443f16005ac" containerName="collect-profiles" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.158372 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a1743fa-87f6-46d8-ad30-fc5b9ea2e761" containerName="registry-server" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.159199 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.172288 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414281-drh6b"] Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.285174 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-fernet-keys\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.285425 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-config-data\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.285534 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-combined-ca-bundle\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.285572 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smpdf\" (UniqueName: \"kubernetes.io/projected/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-kube-api-access-smpdf\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.388878 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-combined-ca-bundle\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.389150 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smpdf\" (UniqueName: \"kubernetes.io/projected/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-kube-api-access-smpdf\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.389385 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-fernet-keys\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.389646 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-config-data\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.394691 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-fernet-keys\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.395012 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-combined-ca-bundle\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.397159 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-config-data\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.407588 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smpdf\" (UniqueName: \"kubernetes.io/projected/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-kube-api-access-smpdf\") pod \"keystone-cron-29414281-drh6b\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.479135 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:00 crc kubenswrapper[4979]: W1204 14:01:00.945413 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55f01a95_3d30_4d74_8cf7_e1b5b35cb6a6.slice/crio-d9fe8b05e60ef9694448e0e25fc830089e748ea51c1de96fec68ee5c39360d43 WatchSource:0}: Error finding container d9fe8b05e60ef9694448e0e25fc830089e748ea51c1de96fec68ee5c39360d43: Status 404 returned error can't find the container with id d9fe8b05e60ef9694448e0e25fc830089e748ea51c1de96fec68ee5c39360d43 Dec 04 14:01:00 crc kubenswrapper[4979]: I1204 14:01:00.950058 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29414281-drh6b"] Dec 04 14:01:01 crc kubenswrapper[4979]: I1204 14:01:01.783600 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414281-drh6b" event={"ID":"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6","Type":"ContainerStarted","Data":"d825b83af179df41f3e16222148baed65432ad657c01877e06779fcf954b3c3e"} Dec 04 14:01:01 crc kubenswrapper[4979]: I1204 14:01:01.783956 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414281-drh6b" event={"ID":"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6","Type":"ContainerStarted","Data":"d9fe8b05e60ef9694448e0e25fc830089e748ea51c1de96fec68ee5c39360d43"} Dec 04 14:01:01 crc kubenswrapper[4979]: I1204 14:01:01.816207 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29414281-drh6b" podStartSLOduration=1.816174836 podStartE2EDuration="1.816174836s" podCreationTimestamp="2025-12-04 14:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:01:01.797776443 +0000 UTC m=+8286.072072267" watchObservedRunningTime="2025-12-04 14:01:01.816174836 +0000 UTC m=+8286.090470640" Dec 04 14:01:07 crc kubenswrapper[4979]: I1204 14:01:07.843512 4979 generic.go:334] "Generic (PLEG): container finished" podID="55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" containerID="d825b83af179df41f3e16222148baed65432ad657c01877e06779fcf954b3c3e" exitCode=0 Dec 04 14:01:07 crc kubenswrapper[4979]: I1204 14:01:07.843609 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414281-drh6b" event={"ID":"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6","Type":"ContainerDied","Data":"d825b83af179df41f3e16222148baed65432ad657c01877e06779fcf954b3c3e"} Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.224054 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.285911 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-combined-ca-bundle\") pod \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.285992 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-fernet-keys\") pod \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.286025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smpdf\" (UniqueName: \"kubernetes.io/projected/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-kube-api-access-smpdf\") pod \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.286187 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-config-data\") pod \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\" (UID: \"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6\") " Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.292761 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-kube-api-access-smpdf" (OuterVolumeSpecName: "kube-api-access-smpdf") pod "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" (UID: "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6"). InnerVolumeSpecName "kube-api-access-smpdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.294662 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" (UID: "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.329309 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" (UID: "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.355261 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-config-data" (OuterVolumeSpecName: "config-data") pod "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" (UID: "55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.389451 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smpdf\" (UniqueName: \"kubernetes.io/projected/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-kube-api-access-smpdf\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.389490 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.389501 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.389514 4979 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.863544 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29414281-drh6b" event={"ID":"55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6","Type":"ContainerDied","Data":"d9fe8b05e60ef9694448e0e25fc830089e748ea51c1de96fec68ee5c39360d43"} Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.863603 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9fe8b05e60ef9694448e0e25fc830089e748ea51c1de96fec68ee5c39360d43" Dec 04 14:01:09 crc kubenswrapper[4979]: I1204 14:01:09.863609 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29414281-drh6b" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.040429 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.040924 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.468558 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jwpv9"] Dec 04 14:02:58 crc kubenswrapper[4979]: E1204 14:02:58.469553 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" containerName="keystone-cron" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.469575 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" containerName="keystone-cron" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.469835 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6" containerName="keystone-cron" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.471808 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.485886 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jwpv9"] Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.627379 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-catalog-content\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.627544 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zn96d\" (UniqueName: \"kubernetes.io/projected/adf7ca5d-79d9-4d51-9358-7e8b380990d7-kube-api-access-zn96d\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.627647 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-utilities\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.729263 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-catalog-content\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.729418 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zn96d\" (UniqueName: \"kubernetes.io/projected/adf7ca5d-79d9-4d51-9358-7e8b380990d7-kube-api-access-zn96d\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.729497 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-utilities\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.729819 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-catalog-content\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.729984 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-utilities\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.751611 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zn96d\" (UniqueName: \"kubernetes.io/projected/adf7ca5d-79d9-4d51-9358-7e8b380990d7-kube-api-access-zn96d\") pod \"redhat-operators-jwpv9\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:58 crc kubenswrapper[4979]: I1204 14:02:58.794814 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:02:59 crc kubenswrapper[4979]: I1204 14:02:59.388964 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jwpv9"] Dec 04 14:03:00 crc kubenswrapper[4979]: I1204 14:03:00.384481 4979 generic.go:334] "Generic (PLEG): container finished" podID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerID="a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147" exitCode=0 Dec 04 14:03:00 crc kubenswrapper[4979]: I1204 14:03:00.384546 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerDied","Data":"a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147"} Dec 04 14:03:00 crc kubenswrapper[4979]: I1204 14:03:00.385718 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerStarted","Data":"fd9bdb072694c11bbd8feddf3efd11bf25ae37bd1e6e8b7709f6a48f387a4294"} Dec 04 14:03:00 crc kubenswrapper[4979]: I1204 14:03:00.387107 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:03:02 crc kubenswrapper[4979]: I1204 14:03:02.405424 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerStarted","Data":"2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb"} Dec 04 14:03:04 crc kubenswrapper[4979]: I1204 14:03:04.439975 4979 generic.go:334] "Generic (PLEG): container finished" podID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerID="2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb" exitCode=0 Dec 04 14:03:04 crc kubenswrapper[4979]: I1204 14:03:04.440094 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerDied","Data":"2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb"} Dec 04 14:03:12 crc kubenswrapper[4979]: I1204 14:03:12.522215 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerStarted","Data":"e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47"} Dec 04 14:03:12 crc kubenswrapper[4979]: I1204 14:03:12.543832 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jwpv9" podStartSLOduration=2.811760036 podStartE2EDuration="14.543809285s" podCreationTimestamp="2025-12-04 14:02:58 +0000 UTC" firstStartedPulling="2025-12-04 14:03:00.386910976 +0000 UTC m=+8404.661206780" lastFinishedPulling="2025-12-04 14:03:12.118960225 +0000 UTC m=+8416.393256029" observedRunningTime="2025-12-04 14:03:12.542930343 +0000 UTC m=+8416.817226167" watchObservedRunningTime="2025-12-04 14:03:12.543809285 +0000 UTC m=+8416.818105089" Dec 04 14:03:18 crc kubenswrapper[4979]: I1204 14:03:18.795483 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:03:18 crc kubenswrapper[4979]: I1204 14:03:18.796128 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:03:18 crc kubenswrapper[4979]: I1204 14:03:18.844426 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:03:19 crc kubenswrapper[4979]: I1204 14:03:19.634553 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:03:19 crc kubenswrapper[4979]: I1204 14:03:19.690341 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jwpv9"] Dec 04 14:03:21 crc kubenswrapper[4979]: I1204 14:03:21.604814 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jwpv9" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="registry-server" containerID="cri-o://e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47" gracePeriod=2 Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.118394 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.180750 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zn96d\" (UniqueName: \"kubernetes.io/projected/adf7ca5d-79d9-4d51-9358-7e8b380990d7-kube-api-access-zn96d\") pod \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.180961 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-catalog-content\") pod \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.181027 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-utilities\") pod \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\" (UID: \"adf7ca5d-79d9-4d51-9358-7e8b380990d7\") " Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.182279 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-utilities" (OuterVolumeSpecName: "utilities") pod "adf7ca5d-79d9-4d51-9358-7e8b380990d7" (UID: "adf7ca5d-79d9-4d51-9358-7e8b380990d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.187067 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf7ca5d-79d9-4d51-9358-7e8b380990d7-kube-api-access-zn96d" (OuterVolumeSpecName: "kube-api-access-zn96d") pod "adf7ca5d-79d9-4d51-9358-7e8b380990d7" (UID: "adf7ca5d-79d9-4d51-9358-7e8b380990d7"). InnerVolumeSpecName "kube-api-access-zn96d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.284494 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zn96d\" (UniqueName: \"kubernetes.io/projected/adf7ca5d-79d9-4d51-9358-7e8b380990d7-kube-api-access-zn96d\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.284528 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.295829 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adf7ca5d-79d9-4d51-9358-7e8b380990d7" (UID: "adf7ca5d-79d9-4d51-9358-7e8b380990d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.386632 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adf7ca5d-79d9-4d51-9358-7e8b380990d7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.617077 4979 generic.go:334] "Generic (PLEG): container finished" podID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerID="e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47" exitCode=0 Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.617123 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerDied","Data":"e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47"} Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.617154 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jwpv9" event={"ID":"adf7ca5d-79d9-4d51-9358-7e8b380990d7","Type":"ContainerDied","Data":"fd9bdb072694c11bbd8feddf3efd11bf25ae37bd1e6e8b7709f6a48f387a4294"} Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.617172 4979 scope.go:117] "RemoveContainer" containerID="e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.617171 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jwpv9" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.647649 4979 scope.go:117] "RemoveContainer" containerID="2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.651890 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jwpv9"] Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.670511 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jwpv9"] Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.674487 4979 scope.go:117] "RemoveContainer" containerID="a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.727003 4979 scope.go:117] "RemoveContainer" containerID="e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47" Dec 04 14:03:22 crc kubenswrapper[4979]: E1204 14:03:22.727602 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47\": container with ID starting with e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47 not found: ID does not exist" containerID="e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.727657 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47"} err="failed to get container status \"e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47\": rpc error: code = NotFound desc = could not find container \"e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47\": container with ID starting with e74c6ae2b0210c42d36df0da0b2752acf62210e93c67928c2d115ad83587ee47 not found: ID does not exist" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.727693 4979 scope.go:117] "RemoveContainer" containerID="2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb" Dec 04 14:03:22 crc kubenswrapper[4979]: E1204 14:03:22.728069 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb\": container with ID starting with 2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb not found: ID does not exist" containerID="2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.728097 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb"} err="failed to get container status \"2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb\": rpc error: code = NotFound desc = could not find container \"2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb\": container with ID starting with 2d433ca6b4caf5a31fade8175215498c50ee4b4ce77fb090c2cc40b90112aacb not found: ID does not exist" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.728114 4979 scope.go:117] "RemoveContainer" containerID="a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147" Dec 04 14:03:22 crc kubenswrapper[4979]: E1204 14:03:22.728731 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147\": container with ID starting with a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147 not found: ID does not exist" containerID="a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147" Dec 04 14:03:22 crc kubenswrapper[4979]: I1204 14:03:22.728792 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147"} err="failed to get container status \"a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147\": rpc error: code = NotFound desc = could not find container \"a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147\": container with ID starting with a06c0a46d47a72eabc39a5a35b5169845e4f788f51fff94096c342cc19dde147 not found: ID does not exist" Dec 04 14:03:24 crc kubenswrapper[4979]: I1204 14:03:24.211904 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" path="/var/lib/kubelet/pods/adf7ca5d-79d9-4d51-9358-7e8b380990d7/volumes" Dec 04 14:03:28 crc kubenswrapper[4979]: I1204 14:03:28.040666 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:03:28 crc kubenswrapper[4979]: I1204 14:03:28.041374 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:03:50 crc kubenswrapper[4979]: I1204 14:03:50.045056 4979 generic.go:334] "Generic (PLEG): container finished" podID="3aacedd4-db12-48a8-9ad0-1c95d92b24f4" containerID="820c6a19defb22b1d5cdb8cfd84dc4b85272c1198dfff58a7d4204a85e7fdbfd" exitCode=0 Dec 04 14:03:50 crc kubenswrapper[4979]: I1204 14:03:50.045663 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" event={"ID":"3aacedd4-db12-48a8-9ad0-1c95d92b24f4","Type":"ContainerDied","Data":"820c6a19defb22b1d5cdb8cfd84dc4b85272c1198dfff58a7d4204a85e7fdbfd"} Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.595118 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.666661 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ssh-key\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.666740 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcr8x\" (UniqueName: \"kubernetes.io/projected/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-kube-api-access-hcr8x\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.666794 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-1\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.666987 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-inventory\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.667117 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-telemetry-combined-ca-bundle\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.667191 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-0\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.667230 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-2\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.667285 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceph\") pod \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\" (UID: \"3aacedd4-db12-48a8-9ad0-1c95d92b24f4\") " Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.672680 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceph" (OuterVolumeSpecName: "ceph") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.686992 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.687054 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-kube-api-access-hcr8x" (OuterVolumeSpecName: "kube-api-access-hcr8x") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "kube-api-access-hcr8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.696887 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.700215 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.707265 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-inventory" (OuterVolumeSpecName: "inventory") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.708230 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.709374 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3aacedd4-db12-48a8-9ad0-1c95d92b24f4" (UID: "3aacedd4-db12-48a8-9ad0-1c95d92b24f4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.771248 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785624 4979 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785648 4979 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785662 4979 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785678 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785693 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785707 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcr8x\" (UniqueName: \"kubernetes.io/projected/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-kube-api-access-hcr8x\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:51 crc kubenswrapper[4979]: I1204 14:03:51.785719 4979 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/3aacedd4-db12-48a8-9ad0-1c95d92b24f4-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.072908 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" event={"ID":"3aacedd4-db12-48a8-9ad0-1c95d92b24f4","Type":"ContainerDied","Data":"8b448fa45abc38d334a7a0a2ffcf01c1c6d11642339c118d00ebd901f1cc1ebd"} Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.072963 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b448fa45abc38d334a7a0a2ffcf01c1c6d11642339c118d00ebd901f1cc1ebd" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.073010 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-openstack-openstack-cell1-szvj8" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.179478 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-h9f9k"] Dec 04 14:03:52 crc kubenswrapper[4979]: E1204 14:03:52.180027 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="registry-server" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.180053 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="registry-server" Dec 04 14:03:52 crc kubenswrapper[4979]: E1204 14:03:52.180092 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aacedd4-db12-48a8-9ad0-1c95d92b24f4" containerName="telemetry-openstack-openstack-cell1" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.180106 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aacedd4-db12-48a8-9ad0-1c95d92b24f4" containerName="telemetry-openstack-openstack-cell1" Dec 04 14:03:52 crc kubenswrapper[4979]: E1204 14:03:52.180144 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="extract-utilities" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.180156 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="extract-utilities" Dec 04 14:03:52 crc kubenswrapper[4979]: E1204 14:03:52.180174 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="extract-content" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.180179 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="extract-content" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.180407 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf7ca5d-79d9-4d51-9358-7e8b380990d7" containerName="registry-server" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.180445 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aacedd4-db12-48a8-9ad0-1c95d92b24f4" containerName="telemetry-openstack-openstack-cell1" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.181214 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.185004 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-sriov-agent-neutron-config" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.185141 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.185183 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.185321 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.185726 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.218357 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-h9f9k"] Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.296777 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.296947 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.297020 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xww6b\" (UniqueName: \"kubernetes.io/projected/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-kube-api-access-xww6b\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.297393 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.297536 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.297695 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.401198 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.401437 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.401600 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.401733 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xww6b\" (UniqueName: \"kubernetes.io/projected/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-kube-api-access-xww6b\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.401831 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.401943 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.406383 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ceph\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.406384 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ssh-key\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.407289 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-inventory\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.407650 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-combined-ca-bundle\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.418562 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-agent-neutron-config-0\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.431178 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xww6b\" (UniqueName: \"kubernetes.io/projected/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-kube-api-access-xww6b\") pod \"neutron-sriov-openstack-openstack-cell1-h9f9k\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:52 crc kubenswrapper[4979]: I1204 14:03:52.515435 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:03:53 crc kubenswrapper[4979]: I1204 14:03:53.167278 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-sriov-openstack-openstack-cell1-h9f9k"] Dec 04 14:03:54 crc kubenswrapper[4979]: I1204 14:03:54.106559 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" event={"ID":"d396b2b7-0262-4e9d-a170-b7cb958b2c3c","Type":"ContainerStarted","Data":"2390e62d64846806e6fe7b70039d8de5f0053688157b92c1fe9c13c1655232b4"} Dec 04 14:03:57 crc kubenswrapper[4979]: I1204 14:03:57.148044 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" event={"ID":"d396b2b7-0262-4e9d-a170-b7cb958b2c3c","Type":"ContainerStarted","Data":"93fe09ef14a5eafc2f1957b3d2f5a94fab5ffd27c16ddf55eda29b591bbd0e46"} Dec 04 14:03:57 crc kubenswrapper[4979]: I1204 14:03:57.171992 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" podStartSLOduration=2.225240898 podStartE2EDuration="5.171965576s" podCreationTimestamp="2025-12-04 14:03:52 +0000 UTC" firstStartedPulling="2025-12-04 14:03:53.170161963 +0000 UTC m=+8457.444457767" lastFinishedPulling="2025-12-04 14:03:56.116886601 +0000 UTC m=+8460.391182445" observedRunningTime="2025-12-04 14:03:57.17172217 +0000 UTC m=+8461.446017974" watchObservedRunningTime="2025-12-04 14:03:57.171965576 +0000 UTC m=+8461.446261420" Dec 04 14:03:58 crc kubenswrapper[4979]: I1204 14:03:58.041224 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:03:58 crc kubenswrapper[4979]: I1204 14:03:58.041343 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:03:58 crc kubenswrapper[4979]: I1204 14:03:58.041398 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 14:03:58 crc kubenswrapper[4979]: I1204 14:03:58.042034 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:03:58 crc kubenswrapper[4979]: I1204 14:03:58.042124 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" gracePeriod=600 Dec 04 14:03:59 crc kubenswrapper[4979]: E1204 14:03:59.109656 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:03:59 crc kubenswrapper[4979]: I1204 14:03:59.179375 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" exitCode=0 Dec 04 14:03:59 crc kubenswrapper[4979]: I1204 14:03:59.179420 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982"} Dec 04 14:03:59 crc kubenswrapper[4979]: I1204 14:03:59.179451 4979 scope.go:117] "RemoveContainer" containerID="33c6921fae4bfde0599067dbe7d0511a70dccbc78375ca402942f8ac5afb854b" Dec 04 14:03:59 crc kubenswrapper[4979]: I1204 14:03:59.180760 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:03:59 crc kubenswrapper[4979]: E1204 14:03:59.181536 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:04:11 crc kubenswrapper[4979]: I1204 14:04:11.198902 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:04:11 crc kubenswrapper[4979]: E1204 14:04:11.199804 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:04:24 crc kubenswrapper[4979]: I1204 14:04:24.201253 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:04:24 crc kubenswrapper[4979]: E1204 14:04:24.203250 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:04:38 crc kubenswrapper[4979]: I1204 14:04:38.199990 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:04:38 crc kubenswrapper[4979]: E1204 14:04:38.200856 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:04:52 crc kubenswrapper[4979]: I1204 14:04:52.199418 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:04:52 crc kubenswrapper[4979]: E1204 14:04:52.200189 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:05:07 crc kubenswrapper[4979]: I1204 14:05:07.199237 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:05:07 crc kubenswrapper[4979]: E1204 14:05:07.200101 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:05:22 crc kubenswrapper[4979]: I1204 14:05:22.199115 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:05:22 crc kubenswrapper[4979]: E1204 14:05:22.200385 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:05:35 crc kubenswrapper[4979]: I1204 14:05:35.199735 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:05:35 crc kubenswrapper[4979]: E1204 14:05:35.200864 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:05:50 crc kubenswrapper[4979]: I1204 14:05:50.200844 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:05:50 crc kubenswrapper[4979]: E1204 14:05:50.201995 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:06:01 crc kubenswrapper[4979]: I1204 14:06:01.199462 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:06:01 crc kubenswrapper[4979]: E1204 14:06:01.200398 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:06:16 crc kubenswrapper[4979]: I1204 14:06:16.217411 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:06:16 crc kubenswrapper[4979]: E1204 14:06:16.218754 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:06:30 crc kubenswrapper[4979]: I1204 14:06:30.198996 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:06:30 crc kubenswrapper[4979]: E1204 14:06:30.199924 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:06:44 crc kubenswrapper[4979]: I1204 14:06:44.200210 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:06:44 crc kubenswrapper[4979]: E1204 14:06:44.201359 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:06:55 crc kubenswrapper[4979]: I1204 14:06:55.200187 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:06:55 crc kubenswrapper[4979]: E1204 14:06:55.201122 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.579540 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7gsqw"] Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.585143 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.669533 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7gsqw"] Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.709935 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-catalog-content\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.710159 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-utilities\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.710294 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tp57h\" (UniqueName: \"kubernetes.io/projected/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-kube-api-access-tp57h\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.811888 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-catalog-content\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.811978 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-utilities\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.812016 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tp57h\" (UniqueName: \"kubernetes.io/projected/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-kube-api-access-tp57h\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.812515 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-catalog-content\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.812575 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-utilities\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.830991 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tp57h\" (UniqueName: \"kubernetes.io/projected/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-kube-api-access-tp57h\") pod \"redhat-marketplace-7gsqw\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:06 crc kubenswrapper[4979]: I1204 14:07:06.905828 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:07 crc kubenswrapper[4979]: I1204 14:07:07.199249 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:07:07 crc kubenswrapper[4979]: E1204 14:07:07.199782 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:07:07 crc kubenswrapper[4979]: I1204 14:07:07.461536 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7gsqw"] Dec 04 14:07:08 crc kubenswrapper[4979]: I1204 14:07:08.235177 4979 generic.go:334] "Generic (PLEG): container finished" podID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerID="07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13" exitCode=0 Dec 04 14:07:08 crc kubenswrapper[4979]: I1204 14:07:08.235239 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7gsqw" event={"ID":"a1b6ffaa-6d79-49c5-8a8d-1b342947af43","Type":"ContainerDied","Data":"07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13"} Dec 04 14:07:08 crc kubenswrapper[4979]: I1204 14:07:08.236587 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7gsqw" event={"ID":"a1b6ffaa-6d79-49c5-8a8d-1b342947af43","Type":"ContainerStarted","Data":"1fbeaad6c2a5465ac263f25fe9cccced9f7917d625d32deb9f422d7fd40cce47"} Dec 04 14:07:10 crc kubenswrapper[4979]: I1204 14:07:10.261739 4979 generic.go:334] "Generic (PLEG): container finished" podID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerID="31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2" exitCode=0 Dec 04 14:07:10 crc kubenswrapper[4979]: I1204 14:07:10.261875 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7gsqw" event={"ID":"a1b6ffaa-6d79-49c5-8a8d-1b342947af43","Type":"ContainerDied","Data":"31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2"} Dec 04 14:07:12 crc kubenswrapper[4979]: I1204 14:07:12.284103 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7gsqw" event={"ID":"a1b6ffaa-6d79-49c5-8a8d-1b342947af43","Type":"ContainerStarted","Data":"b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725"} Dec 04 14:07:12 crc kubenswrapper[4979]: I1204 14:07:12.311197 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7gsqw" podStartSLOduration=3.383874868 podStartE2EDuration="6.311170235s" podCreationTimestamp="2025-12-04 14:07:06 +0000 UTC" firstStartedPulling="2025-12-04 14:07:08.238696225 +0000 UTC m=+8652.512992029" lastFinishedPulling="2025-12-04 14:07:11.165991592 +0000 UTC m=+8655.440287396" observedRunningTime="2025-12-04 14:07:12.304676964 +0000 UTC m=+8656.578972768" watchObservedRunningTime="2025-12-04 14:07:12.311170235 +0000 UTC m=+8656.585466039" Dec 04 14:07:14 crc kubenswrapper[4979]: I1204 14:07:14.319145 4979 generic.go:334] "Generic (PLEG): container finished" podID="d396b2b7-0262-4e9d-a170-b7cb958b2c3c" containerID="93fe09ef14a5eafc2f1957b3d2f5a94fab5ffd27c16ddf55eda29b591bbd0e46" exitCode=0 Dec 04 14:07:14 crc kubenswrapper[4979]: I1204 14:07:14.319238 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" event={"ID":"d396b2b7-0262-4e9d-a170-b7cb958b2c3c","Type":"ContainerDied","Data":"93fe09ef14a5eafc2f1957b3d2f5a94fab5ffd27c16ddf55eda29b591bbd0e46"} Dec 04 14:07:15 crc kubenswrapper[4979]: I1204 14:07:15.856411 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.048654 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xww6b\" (UniqueName: \"kubernetes.io/projected/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-kube-api-access-xww6b\") pod \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.048788 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ceph\") pod \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.048838 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-agent-neutron-config-0\") pod \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.048891 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ssh-key\") pod \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.048955 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-combined-ca-bundle\") pod \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.049083 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-inventory\") pod \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\" (UID: \"d396b2b7-0262-4e9d-a170-b7cb958b2c3c\") " Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.055115 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ceph" (OuterVolumeSpecName: "ceph") pod "d396b2b7-0262-4e9d-a170-b7cb958b2c3c" (UID: "d396b2b7-0262-4e9d-a170-b7cb958b2c3c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.055661 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-combined-ca-bundle" (OuterVolumeSpecName: "neutron-sriov-combined-ca-bundle") pod "d396b2b7-0262-4e9d-a170-b7cb958b2c3c" (UID: "d396b2b7-0262-4e9d-a170-b7cb958b2c3c"). InnerVolumeSpecName "neutron-sriov-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.057451 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-kube-api-access-xww6b" (OuterVolumeSpecName: "kube-api-access-xww6b") pod "d396b2b7-0262-4e9d-a170-b7cb958b2c3c" (UID: "d396b2b7-0262-4e9d-a170-b7cb958b2c3c"). InnerVolumeSpecName "kube-api-access-xww6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.080770 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d396b2b7-0262-4e9d-a170-b7cb958b2c3c" (UID: "d396b2b7-0262-4e9d-a170-b7cb958b2c3c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.083245 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-inventory" (OuterVolumeSpecName: "inventory") pod "d396b2b7-0262-4e9d-a170-b7cb958b2c3c" (UID: "d396b2b7-0262-4e9d-a170-b7cb958b2c3c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.083914 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-sriov-agent-neutron-config-0") pod "d396b2b7-0262-4e9d-a170-b7cb958b2c3c" (UID: "d396b2b7-0262-4e9d-a170-b7cb958b2c3c"). InnerVolumeSpecName "neutron-sriov-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.151001 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.151032 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xww6b\" (UniqueName: \"kubernetes.io/projected/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-kube-api-access-xww6b\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.151043 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.151055 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.151064 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.151073 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-sriov-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d396b2b7-0262-4e9d-a170-b7cb958b2c3c-neutron-sriov-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:16 crc kubenswrapper[4979]: E1204 14:07:16.339206 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd396b2b7_0262_4e9d_a170_b7cb958b2c3c.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd396b2b7_0262_4e9d_a170_b7cb958b2c3c.slice/crio-2390e62d64846806e6fe7b70039d8de5f0053688157b92c1fe9c13c1655232b4\": RecentStats: unable to find data in memory cache]" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.342140 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" event={"ID":"d396b2b7-0262-4e9d-a170-b7cb958b2c3c","Type":"ContainerDied","Data":"2390e62d64846806e6fe7b70039d8de5f0053688157b92c1fe9c13c1655232b4"} Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.342184 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2390e62d64846806e6fe7b70039d8de5f0053688157b92c1fe9c13c1655232b4" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.342238 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-sriov-openstack-openstack-cell1-h9f9k" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.442870 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6"] Dec 04 14:07:16 crc kubenswrapper[4979]: E1204 14:07:16.443635 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d396b2b7-0262-4e9d-a170-b7cb958b2c3c" containerName="neutron-sriov-openstack-openstack-cell1" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.443648 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="d396b2b7-0262-4e9d-a170-b7cb958b2c3c" containerName="neutron-sriov-openstack-openstack-cell1" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.443873 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="d396b2b7-0262-4e9d-a170-b7cb958b2c3c" containerName="neutron-sriov-openstack-openstack-cell1" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.444700 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.447881 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.450599 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.450690 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-dhcp-agent-neutron-config" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.451211 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.451217 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.482902 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6"] Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.561735 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.561815 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.562176 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.562247 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.562558 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.562690 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqxk2\" (UniqueName: \"kubernetes.io/projected/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-kube-api-access-sqxk2\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.664449 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.664551 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqxk2\" (UniqueName: \"kubernetes.io/projected/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-kube-api-access-sqxk2\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.664657 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.664706 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.664791 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.664859 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.671169 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-agent-neutron-config-0\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.672013 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-inventory\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.672516 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ssh-key\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.673544 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ceph\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.676225 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-combined-ca-bundle\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.694711 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqxk2\" (UniqueName: \"kubernetes.io/projected/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-kube-api-access-sqxk2\") pod \"neutron-dhcp-openstack-openstack-cell1-2rrr6\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.783377 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.907567 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.908249 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:16 crc kubenswrapper[4979]: I1204 14:07:16.985591 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:17 crc kubenswrapper[4979]: I1204 14:07:17.406321 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:17 crc kubenswrapper[4979]: I1204 14:07:17.433508 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6"] Dec 04 14:07:17 crc kubenswrapper[4979]: I1204 14:07:17.471681 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7gsqw"] Dec 04 14:07:18 crc kubenswrapper[4979]: I1204 14:07:18.364358 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" event={"ID":"7b19b05f-caa7-45f7-9eb5-890ef8fecaae","Type":"ContainerStarted","Data":"a0ce6316c0bde444a3ab21d3abec3b885005a1c426de79807841928523bc9286"} Dec 04 14:07:18 crc kubenswrapper[4979]: I1204 14:07:18.365004 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" event={"ID":"7b19b05f-caa7-45f7-9eb5-890ef8fecaae","Type":"ContainerStarted","Data":"040601427d4e3722ba8525c592b738b9c4126bd32e70fcf4dc4ff8a89fe277bb"} Dec 04 14:07:18 crc kubenswrapper[4979]: I1204 14:07:18.383514 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" podStartSLOduration=1.961335179 podStartE2EDuration="2.383486789s" podCreationTimestamp="2025-12-04 14:07:16 +0000 UTC" firstStartedPulling="2025-12-04 14:07:17.435859535 +0000 UTC m=+8661.710155339" lastFinishedPulling="2025-12-04 14:07:17.858011125 +0000 UTC m=+8662.132306949" observedRunningTime="2025-12-04 14:07:18.3804675 +0000 UTC m=+8662.654763324" watchObservedRunningTime="2025-12-04 14:07:18.383486789 +0000 UTC m=+8662.657782593" Dec 04 14:07:19 crc kubenswrapper[4979]: I1204 14:07:19.375043 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7gsqw" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="registry-server" containerID="cri-o://b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725" gracePeriod=2 Dec 04 14:07:19 crc kubenswrapper[4979]: I1204 14:07:19.880339 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.038734 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-utilities\") pod \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.038827 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tp57h\" (UniqueName: \"kubernetes.io/projected/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-kube-api-access-tp57h\") pod \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.038944 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-catalog-content\") pod \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\" (UID: \"a1b6ffaa-6d79-49c5-8a8d-1b342947af43\") " Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.040143 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-utilities" (OuterVolumeSpecName: "utilities") pod "a1b6ffaa-6d79-49c5-8a8d-1b342947af43" (UID: "a1b6ffaa-6d79-49c5-8a8d-1b342947af43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.044904 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-kube-api-access-tp57h" (OuterVolumeSpecName: "kube-api-access-tp57h") pod "a1b6ffaa-6d79-49c5-8a8d-1b342947af43" (UID: "a1b6ffaa-6d79-49c5-8a8d-1b342947af43"). InnerVolumeSpecName "kube-api-access-tp57h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.060372 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1b6ffaa-6d79-49c5-8a8d-1b342947af43" (UID: "a1b6ffaa-6d79-49c5-8a8d-1b342947af43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.141869 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.141907 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tp57h\" (UniqueName: \"kubernetes.io/projected/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-kube-api-access-tp57h\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.141919 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1b6ffaa-6d79-49c5-8a8d-1b342947af43-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.199894 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:07:20 crc kubenswrapper[4979]: E1204 14:07:20.200232 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.401444 4979 generic.go:334] "Generic (PLEG): container finished" podID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerID="b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725" exitCode=0 Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.401483 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7gsqw" event={"ID":"a1b6ffaa-6d79-49c5-8a8d-1b342947af43","Type":"ContainerDied","Data":"b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725"} Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.401507 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7gsqw" event={"ID":"a1b6ffaa-6d79-49c5-8a8d-1b342947af43","Type":"ContainerDied","Data":"1fbeaad6c2a5465ac263f25fe9cccced9f7917d625d32deb9f422d7fd40cce47"} Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.401513 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7gsqw" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.401522 4979 scope.go:117] "RemoveContainer" containerID="b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.433089 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7gsqw"] Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.439317 4979 scope.go:117] "RemoveContainer" containerID="31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.446668 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7gsqw"] Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.558811 4979 scope.go:117] "RemoveContainer" containerID="07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.612215 4979 scope.go:117] "RemoveContainer" containerID="b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725" Dec 04 14:07:20 crc kubenswrapper[4979]: E1204 14:07:20.639567 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725\": container with ID starting with b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725 not found: ID does not exist" containerID="b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.639624 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725"} err="failed to get container status \"b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725\": rpc error: code = NotFound desc = could not find container \"b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725\": container with ID starting with b1808dbb74fac6d67c416743665f6fe282ecb78247912822d28b1dbef671a725 not found: ID does not exist" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.639673 4979 scope.go:117] "RemoveContainer" containerID="31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2" Dec 04 14:07:20 crc kubenswrapper[4979]: E1204 14:07:20.640814 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2\": container with ID starting with 31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2 not found: ID does not exist" containerID="31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.640847 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2"} err="failed to get container status \"31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2\": rpc error: code = NotFound desc = could not find container \"31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2\": container with ID starting with 31666766f58354e98c2460255f29917282b08af0dbae2f8459e642bee22e40c2 not found: ID does not exist" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.640865 4979 scope.go:117] "RemoveContainer" containerID="07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13" Dec 04 14:07:20 crc kubenswrapper[4979]: E1204 14:07:20.641313 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13\": container with ID starting with 07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13 not found: ID does not exist" containerID="07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13" Dec 04 14:07:20 crc kubenswrapper[4979]: I1204 14:07:20.641335 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13"} err="failed to get container status \"07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13\": rpc error: code = NotFound desc = could not find container \"07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13\": container with ID starting with 07dc93b8694d9e7a1e5f90bce03e5e4fa152747a683cd7913299a7d9898cbd13 not found: ID does not exist" Dec 04 14:07:22 crc kubenswrapper[4979]: I1204 14:07:22.214151 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" path="/var/lib/kubelet/pods/a1b6ffaa-6d79-49c5-8a8d-1b342947af43/volumes" Dec 04 14:07:32 crc kubenswrapper[4979]: I1204 14:07:32.201280 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:07:32 crc kubenswrapper[4979]: E1204 14:07:32.202505 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:07:45 crc kubenswrapper[4979]: I1204 14:07:45.199764 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:07:45 crc kubenswrapper[4979]: E1204 14:07:45.200656 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:07:58 crc kubenswrapper[4979]: I1204 14:07:58.199285 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:07:58 crc kubenswrapper[4979]: E1204 14:07:58.200069 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:08:09 crc kubenswrapper[4979]: I1204 14:08:09.200535 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:08:09 crc kubenswrapper[4979]: E1204 14:08:09.204056 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:08:22 crc kubenswrapper[4979]: I1204 14:08:22.199164 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:08:22 crc kubenswrapper[4979]: E1204 14:08:22.200106 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:08:35 crc kubenswrapper[4979]: I1204 14:08:35.198926 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:08:35 crc kubenswrapper[4979]: E1204 14:08:35.199881 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:08:50 crc kubenswrapper[4979]: I1204 14:08:50.199819 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:08:50 crc kubenswrapper[4979]: E1204 14:08:50.201852 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:09:04 crc kubenswrapper[4979]: I1204 14:09:04.199443 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:09:04 crc kubenswrapper[4979]: I1204 14:09:04.567472 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"2dda8f444bccd60d9a9921986f15443592d20a8839d0aa1af7274bb26ac30114"} Dec 04 14:10:37 crc kubenswrapper[4979]: I1204 14:10:37.781552 4979 generic.go:334] "Generic (PLEG): container finished" podID="7b19b05f-caa7-45f7-9eb5-890ef8fecaae" containerID="a0ce6316c0bde444a3ab21d3abec3b885005a1c426de79807841928523bc9286" exitCode=0 Dec 04 14:10:37 crc kubenswrapper[4979]: I1204 14:10:37.781659 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" event={"ID":"7b19b05f-caa7-45f7-9eb5-890ef8fecaae","Type":"ContainerDied","Data":"a0ce6316c0bde444a3ab21d3abec3b885005a1c426de79807841928523bc9286"} Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.269019 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.465015 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ceph\") pod \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.465364 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-agent-neutron-config-0\") pod \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.465410 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqxk2\" (UniqueName: \"kubernetes.io/projected/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-kube-api-access-sqxk2\") pod \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.465485 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-combined-ca-bundle\") pod \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.465524 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ssh-key\") pod \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.465543 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-inventory\") pod \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\" (UID: \"7b19b05f-caa7-45f7-9eb5-890ef8fecaae\") " Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.490719 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-combined-ca-bundle" (OuterVolumeSpecName: "neutron-dhcp-combined-ca-bundle") pod "7b19b05f-caa7-45f7-9eb5-890ef8fecaae" (UID: "7b19b05f-caa7-45f7-9eb5-890ef8fecaae"). InnerVolumeSpecName "neutron-dhcp-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.495238 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-kube-api-access-sqxk2" (OuterVolumeSpecName: "kube-api-access-sqxk2") pod "7b19b05f-caa7-45f7-9eb5-890ef8fecaae" (UID: "7b19b05f-caa7-45f7-9eb5-890ef8fecaae"). InnerVolumeSpecName "kube-api-access-sqxk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.510515 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ceph" (OuterVolumeSpecName: "ceph") pod "7b19b05f-caa7-45f7-9eb5-890ef8fecaae" (UID: "7b19b05f-caa7-45f7-9eb5-890ef8fecaae"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.565383 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-inventory" (OuterVolumeSpecName: "inventory") pod "7b19b05f-caa7-45f7-9eb5-890ef8fecaae" (UID: "7b19b05f-caa7-45f7-9eb5-890ef8fecaae"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.566966 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7b19b05f-caa7-45f7-9eb5-890ef8fecaae" (UID: "7b19b05f-caa7-45f7-9eb5-890ef8fecaae"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.567584 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqxk2\" (UniqueName: \"kubernetes.io/projected/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-kube-api-access-sqxk2\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.567604 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.567617 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.567673 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.567682 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.571293 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-dhcp-agent-neutron-config-0") pod "7b19b05f-caa7-45f7-9eb5-890ef8fecaae" (UID: "7b19b05f-caa7-45f7-9eb5-890ef8fecaae"). InnerVolumeSpecName "neutron-dhcp-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.670241 4979 reconciler_common.go:293] "Volume detached for volume \"neutron-dhcp-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/7b19b05f-caa7-45f7-9eb5-890ef8fecaae-neutron-dhcp-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.801926 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" event={"ID":"7b19b05f-caa7-45f7-9eb5-890ef8fecaae","Type":"ContainerDied","Data":"040601427d4e3722ba8525c592b738b9c4126bd32e70fcf4dc4ff8a89fe277bb"} Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.801991 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="040601427d4e3722ba8525c592b738b9c4126bd32e70fcf4dc4ff8a89fe277bb" Dec 04 14:10:39 crc kubenswrapper[4979]: I1204 14:10:39.802001 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-dhcp-openstack-openstack-cell1-2rrr6" Dec 04 14:11:07 crc kubenswrapper[4979]: I1204 14:11:07.180559 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:11:07 crc kubenswrapper[4979]: I1204 14:11:07.181196 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" containerName="nova-cell0-conductor-conductor" containerID="cri-o://2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e" gracePeriod=30 Dec 04 14:11:07 crc kubenswrapper[4979]: I1204 14:11:07.208523 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:11:07 crc kubenswrapper[4979]: I1204 14:11:07.208776 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-conductor-0" podUID="235f670f-55af-4667-998a-9d0066bc01d1" containerName="nova-cell1-conductor-conductor" containerID="cri-o://1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8" gracePeriod=30 Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.345385 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.346045 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-log" containerID="cri-o://e15b275173b8bd8e14ab47413fbf97633b85db28701be0e8cbdb009087d43c9f" gracePeriod=30 Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.346559 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-api" containerID="cri-o://ce04a68a8d5882a2a8474677b476c012dad476be2cac35c4de03abbf3a0a803b" gracePeriod=30 Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.370485 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.370732 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="cf1145a5-ea36-490c-9360-98feed19a052" containerName="nova-scheduler-scheduler" containerID="cri-o://642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" gracePeriod=30 Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.385978 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.386248 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-log" containerID="cri-o://9f1eda5655293e5c52af5fb24746e23eec2651f8e849d13062d637d2cf6d7f81" gracePeriod=30 Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.386438 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-metadata" containerID="cri-o://d0a361d3f246a9a0b46fcb15732158eba4b629166f2aa067d3ab69b289653692" gracePeriod=30 Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.682873 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.721844 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58gsl\" (UniqueName: \"kubernetes.io/projected/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-kube-api-access-58gsl\") pod \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.721948 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-combined-ca-bundle\") pod \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.721975 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-config-data\") pod \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\" (UID: \"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458\") " Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.744614 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-kube-api-access-58gsl" (OuterVolumeSpecName: "kube-api-access-58gsl") pod "b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" (UID: "b4f223cc-e5ec-4e34-ba2d-44d8a47ad458"). InnerVolumeSpecName "kube-api-access-58gsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.777584 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" (UID: "b4f223cc-e5ec-4e34-ba2d-44d8a47ad458"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.826633 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.826681 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58gsl\" (UniqueName: \"kubernetes.io/projected/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-kube-api-access-58gsl\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.827662 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.832466 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-config-data" (OuterVolumeSpecName: "config-data") pod "b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" (UID: "b4f223cc-e5ec-4e34-ba2d-44d8a47ad458"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.928041 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-config-data\") pod \"235f670f-55af-4667-998a-9d0066bc01d1\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.928157 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-combined-ca-bundle\") pod \"235f670f-55af-4667-998a-9d0066bc01d1\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.928232 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgm2r\" (UniqueName: \"kubernetes.io/projected/235f670f-55af-4667-998a-9d0066bc01d1-kube-api-access-lgm2r\") pod \"235f670f-55af-4667-998a-9d0066bc01d1\" (UID: \"235f670f-55af-4667-998a-9d0066bc01d1\") " Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.928713 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:08 crc kubenswrapper[4979]: I1204 14:11:08.933742 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/235f670f-55af-4667-998a-9d0066bc01d1-kube-api-access-lgm2r" (OuterVolumeSpecName: "kube-api-access-lgm2r") pod "235f670f-55af-4667-998a-9d0066bc01d1" (UID: "235f670f-55af-4667-998a-9d0066bc01d1"). InnerVolumeSpecName "kube-api-access-lgm2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.016818 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-config-data" (OuterVolumeSpecName: "config-data") pod "235f670f-55af-4667-998a-9d0066bc01d1" (UID: "235f670f-55af-4667-998a-9d0066bc01d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.025428 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "235f670f-55af-4667-998a-9d0066bc01d1" (UID: "235f670f-55af-4667-998a-9d0066bc01d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.030957 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.030985 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/235f670f-55af-4667-998a-9d0066bc01d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.030996 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgm2r\" (UniqueName: \"kubernetes.io/projected/235f670f-55af-4667-998a-9d0066bc01d1-kube-api-access-lgm2r\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.141780 4979 generic.go:334] "Generic (PLEG): container finished" podID="88d88d25-879c-461c-a146-89bc738ef3ff" containerID="e15b275173b8bd8e14ab47413fbf97633b85db28701be0e8cbdb009087d43c9f" exitCode=143 Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.142068 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88d88d25-879c-461c-a146-89bc738ef3ff","Type":"ContainerDied","Data":"e15b275173b8bd8e14ab47413fbf97633b85db28701be0e8cbdb009087d43c9f"} Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.143951 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerID="9f1eda5655293e5c52af5fb24746e23eec2651f8e849d13062d637d2cf6d7f81" exitCode=143 Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.144106 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc6d010c-f547-4ea1-9d88-16a8a85834bf","Type":"ContainerDied","Data":"9f1eda5655293e5c52af5fb24746e23eec2651f8e849d13062d637d2cf6d7f81"} Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.145673 4979 generic.go:334] "Generic (PLEG): container finished" podID="b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" containerID="2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e" exitCode=0 Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.145811 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458","Type":"ContainerDied","Data":"2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e"} Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.145908 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"b4f223cc-e5ec-4e34-ba2d-44d8a47ad458","Type":"ContainerDied","Data":"0b2e18202413d32c534a7833e482c549758aeba1f849bb4345fdfac63af74be5"} Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.146016 4979 scope.go:117] "RemoveContainer" containerID="2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.146241 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.148187 4979 generic.go:334] "Generic (PLEG): container finished" podID="235f670f-55af-4667-998a-9d0066bc01d1" containerID="1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8" exitCode=0 Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.148223 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.148231 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"235f670f-55af-4667-998a-9d0066bc01d1","Type":"ContainerDied","Data":"1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8"} Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.148258 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"235f670f-55af-4667-998a-9d0066bc01d1","Type":"ContainerDied","Data":"333398ff21ea6229bcc352dee8bd6343d092c0921c76c3e6c5cdf1e68245ff56"} Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.174998 4979 scope.go:117] "RemoveContainer" containerID="2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.175524 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e\": container with ID starting with 2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e not found: ID does not exist" containerID="2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.175558 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e"} err="failed to get container status \"2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e\": rpc error: code = NotFound desc = could not find container \"2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e\": container with ID starting with 2b030ae10727993bd63c04274125549bf8e60e90b171ca71cbc43c784a00a64e not found: ID does not exist" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.175579 4979 scope.go:117] "RemoveContainer" containerID="1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.199256 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.229447 4979 scope.go:117] "RemoveContainer" containerID="1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.229858 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8\": container with ID starting with 1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8 not found: ID does not exist" containerID="1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.229889 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8"} err="failed to get container status \"1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8\": rpc error: code = NotFound desc = could not find container \"1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8\": container with ID starting with 1d2ef9fe9766c28809724b5488c2fb7e562fdea9703779c70a98f5403c1979b8 not found: ID does not exist" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.231202 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.250814 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.261349 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.273464 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.273968 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b19b05f-caa7-45f7-9eb5-890ef8fecaae" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.273995 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b19b05f-caa7-45f7-9eb5-890ef8fecaae" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.274028 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="registry-server" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274034 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="registry-server" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.274052 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="extract-utilities" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274058 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="extract-utilities" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.274074 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="235f670f-55af-4667-998a-9d0066bc01d1" containerName="nova-cell1-conductor-conductor" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274079 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="235f670f-55af-4667-998a-9d0066bc01d1" containerName="nova-cell1-conductor-conductor" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.274086 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="extract-content" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274091 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="extract-content" Dec 04 14:11:09 crc kubenswrapper[4979]: E1204 14:11:09.274101 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" containerName="nova-cell0-conductor-conductor" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274107 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" containerName="nova-cell0-conductor-conductor" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274290 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="235f670f-55af-4667-998a-9d0066bc01d1" containerName="nova-cell1-conductor-conductor" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274319 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1b6ffaa-6d79-49c5-8a8d-1b342947af43" containerName="registry-server" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274336 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" containerName="nova-cell0-conductor-conductor" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.274352 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b19b05f-caa7-45f7-9eb5-890ef8fecaae" containerName="neutron-dhcp-openstack-openstack-cell1" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.275059 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.279192 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.286575 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.288476 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.291087 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.300042 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.318070 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.344355 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d082d430-346e-43ac-adc2-cfce3ff1d996-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.344461 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.344486 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftng9\" (UniqueName: \"kubernetes.io/projected/d082d430-346e-43ac-adc2-cfce3ff1d996-kube-api-access-ftng9\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.344561 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.344581 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d082d430-346e-43ac-adc2-cfce3ff1d996-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.344595 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx4j5\" (UniqueName: \"kubernetes.io/projected/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-kube-api-access-hx4j5\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.446186 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.446241 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftng9\" (UniqueName: \"kubernetes.io/projected/d082d430-346e-43ac-adc2-cfce3ff1d996-kube-api-access-ftng9\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.446357 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.446384 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d082d430-346e-43ac-adc2-cfce3ff1d996-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.446404 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx4j5\" (UniqueName: \"kubernetes.io/projected/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-kube-api-access-hx4j5\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.446501 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d082d430-346e-43ac-adc2-cfce3ff1d996-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.457576 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.457780 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d082d430-346e-43ac-adc2-cfce3ff1d996-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.457822 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.458296 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d082d430-346e-43ac-adc2-cfce3ff1d996-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.467773 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftng9\" (UniqueName: \"kubernetes.io/projected/d082d430-346e-43ac-adc2-cfce3ff1d996-kube-api-access-ftng9\") pod \"nova-cell1-conductor-0\" (UID: \"d082d430-346e-43ac-adc2-cfce3ff1d996\") " pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.469425 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx4j5\" (UniqueName: \"kubernetes.io/projected/1e0ffe92-c44a-454c-838f-d53f10f4a5f4-kube-api-access-hx4j5\") pod \"nova-cell0-conductor-0\" (UID: \"1e0ffe92-c44a-454c-838f-d53f10f4a5f4\") " pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.615184 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:09 crc kubenswrapper[4979]: I1204 14:11:09.622221 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:10 crc kubenswrapper[4979]: I1204 14:11:10.184705 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 04 14:11:10 crc kubenswrapper[4979]: I1204 14:11:10.194101 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 04 14:11:10 crc kubenswrapper[4979]: I1204 14:11:10.216226 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="235f670f-55af-4667-998a-9d0066bc01d1" path="/var/lib/kubelet/pods/235f670f-55af-4667-998a-9d0066bc01d1/volumes" Dec 04 14:11:10 crc kubenswrapper[4979]: I1204 14:11:10.216832 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4f223cc-e5ec-4e34-ba2d-44d8a47ad458" path="/var/lib/kubelet/pods/b4f223cc-e5ec-4e34-ba2d-44d8a47ad458/volumes" Dec 04 14:11:10 crc kubenswrapper[4979]: E1204 14:11:10.836383 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:11:10 crc kubenswrapper[4979]: E1204 14:11:10.838991 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:11:10 crc kubenswrapper[4979]: E1204 14:11:10.840991 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:11:10 crc kubenswrapper[4979]: E1204 14:11:10.841044 4979 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cf1145a5-ea36-490c-9360-98feed19a052" containerName="nova-scheduler-scheduler" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.186167 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1e0ffe92-c44a-454c-838f-d53f10f4a5f4","Type":"ContainerStarted","Data":"d64ff638f237a2abc99c2a154e8507ce4b4823cb08d6c148641ae0711452f131"} Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.186214 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"1e0ffe92-c44a-454c-838f-d53f10f4a5f4","Type":"ContainerStarted","Data":"9f3941682127e9b80402d600be4a39f454733c46965a70c92a2b3f1fc871b426"} Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.186264 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.187598 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d082d430-346e-43ac-adc2-cfce3ff1d996","Type":"ContainerStarted","Data":"0f4f27a083474fc8e6ce1dbea820c6ce91098c1fa6737a448eae560ef83adea5"} Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.187647 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"d082d430-346e-43ac-adc2-cfce3ff1d996","Type":"ContainerStarted","Data":"c32f0c092a939fe9fa8058e89c16c5b36d5d55d1b152d18f6bdb51a60e8cceee"} Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.187780 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.208688 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.208672331 podStartE2EDuration="2.208672331s" podCreationTimestamp="2025-12-04 14:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:11.206504124 +0000 UTC m=+8895.480799948" watchObservedRunningTime="2025-12-04 14:11:11.208672331 +0000 UTC m=+8895.482968135" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.238925 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.238901565 podStartE2EDuration="2.238901565s" podCreationTimestamp="2025-12-04 14:11:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:11.235695971 +0000 UTC m=+8895.509991765" watchObservedRunningTime="2025-12-04 14:11:11.238901565 +0000 UTC m=+8895.513197369" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.845228 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": read tcp 10.217.0.2:38092->10.217.1.87:8774: read: connection reset by peer" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.845257 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-api-0" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.87:8774/\": read tcp 10.217.0.2:38088->10.217.1.87:8774: read: connection reset by peer" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.997757 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.86:8775/\": read tcp 10.217.0.2:40560->10.217.1.86:8775: read: connection reset by peer" Dec 04 14:11:11 crc kubenswrapper[4979]: I1204 14:11:11.998073 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.86:8775/\": read tcp 10.217.0.2:40556->10.217.1.86:8775: read: connection reset by peer" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.197492 4979 generic.go:334] "Generic (PLEG): container finished" podID="88d88d25-879c-461c-a146-89bc738ef3ff" containerID="ce04a68a8d5882a2a8474677b476c012dad476be2cac35c4de03abbf3a0a803b" exitCode=0 Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.197553 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88d88d25-879c-461c-a146-89bc738ef3ff","Type":"ContainerDied","Data":"ce04a68a8d5882a2a8474677b476c012dad476be2cac35c4de03abbf3a0a803b"} Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.201782 4979 generic.go:334] "Generic (PLEG): container finished" podID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerID="d0a361d3f246a9a0b46fcb15732158eba4b629166f2aa067d3ab69b289653692" exitCode=0 Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.219464 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc6d010c-f547-4ea1-9d88-16a8a85834bf","Type":"ContainerDied","Data":"d0a361d3f246a9a0b46fcb15732158eba4b629166f2aa067d3ab69b289653692"} Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.579610 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.622385 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc6d010c-f547-4ea1-9d88-16a8a85834bf-logs\") pod \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.622448 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-config-data\") pod \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.622621 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgh8j\" (UniqueName: \"kubernetes.io/projected/bc6d010c-f547-4ea1-9d88-16a8a85834bf-kube-api-access-cgh8j\") pod \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.622678 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-combined-ca-bundle\") pod \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\" (UID: \"bc6d010c-f547-4ea1-9d88-16a8a85834bf\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.627605 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc6d010c-f547-4ea1-9d88-16a8a85834bf-logs" (OuterVolumeSpecName: "logs") pod "bc6d010c-f547-4ea1-9d88-16a8a85834bf" (UID: "bc6d010c-f547-4ea1-9d88-16a8a85834bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.664573 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-config-data" (OuterVolumeSpecName: "config-data") pod "bc6d010c-f547-4ea1-9d88-16a8a85834bf" (UID: "bc6d010c-f547-4ea1-9d88-16a8a85834bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.666468 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc6d010c-f547-4ea1-9d88-16a8a85834bf-kube-api-access-cgh8j" (OuterVolumeSpecName: "kube-api-access-cgh8j") pod "bc6d010c-f547-4ea1-9d88-16a8a85834bf" (UID: "bc6d010c-f547-4ea1-9d88-16a8a85834bf"). InnerVolumeSpecName "kube-api-access-cgh8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.666577 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc6d010c-f547-4ea1-9d88-16a8a85834bf" (UID: "bc6d010c-f547-4ea1-9d88-16a8a85834bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.725731 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.725774 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bc6d010c-f547-4ea1-9d88-16a8a85834bf-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.725790 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc6d010c-f547-4ea1-9d88-16a8a85834bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.725804 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgh8j\" (UniqueName: \"kubernetes.io/projected/bc6d010c-f547-4ea1-9d88-16a8a85834bf-kube-api-access-cgh8j\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.752339 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.827244 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-combined-ca-bundle\") pod \"88d88d25-879c-461c-a146-89bc738ef3ff\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.827418 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d88d25-879c-461c-a146-89bc738ef3ff-logs\") pod \"88d88d25-879c-461c-a146-89bc738ef3ff\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.827503 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqp64\" (UniqueName: \"kubernetes.io/projected/88d88d25-879c-461c-a146-89bc738ef3ff-kube-api-access-rqp64\") pod \"88d88d25-879c-461c-a146-89bc738ef3ff\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.827537 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-config-data\") pod \"88d88d25-879c-461c-a146-89bc738ef3ff\" (UID: \"88d88d25-879c-461c-a146-89bc738ef3ff\") " Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.829387 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88d88d25-879c-461c-a146-89bc738ef3ff-logs" (OuterVolumeSpecName: "logs") pod "88d88d25-879c-461c-a146-89bc738ef3ff" (UID: "88d88d25-879c-461c-a146-89bc738ef3ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.830780 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts"] Dec 04 14:11:12 crc kubenswrapper[4979]: E1204 14:11:12.831335 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-metadata" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.831357 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-metadata" Dec 04 14:11:12 crc kubenswrapper[4979]: E1204 14:11:12.831427 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-api" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.831437 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-api" Dec 04 14:11:12 crc kubenswrapper[4979]: E1204 14:11:12.831446 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-log" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.831455 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-log" Dec 04 14:11:12 crc kubenswrapper[4979]: E1204 14:11:12.831476 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-log" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.831484 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-log" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.833494 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-log" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.833522 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-log" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.833552 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" containerName="nova-api-api" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.833562 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" containerName="nova-metadata-metadata" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.834605 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.837606 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88d88d25-879c-461c-a146-89bc738ef3ff-kube-api-access-rqp64" (OuterVolumeSpecName: "kube-api-access-rqp64") pod "88d88d25-879c-461c-a146-89bc738ef3ff" (UID: "88d88d25-879c-461c-a146-89bc738ef3ff"). InnerVolumeSpecName "kube-api-access-rqp64". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.839113 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.840930 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.841228 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-cells-global-config" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.841430 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-adoption-secret" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.841661 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-dockercfg-m8zp5" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.841816 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.845398 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-cell1" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.853120 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts"] Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.923545 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-config-data" (OuterVolumeSpecName: "config-data") pod "88d88d25-879c-461c-a146-89bc738ef3ff" (UID: "88d88d25-879c-461c-a146-89bc738ef3ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.925120 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88d88d25-879c-461c-a146-89bc738ef3ff" (UID: "88d88d25-879c-461c-a146-89bc738ef3ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.929866 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpbfq\" (UniqueName: \"kubernetes.io/projected/351d7f45-a24f-46da-a9c9-2b598fff069d-kube-api-access-lpbfq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.929918 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.929952 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.929981 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.929999 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930027 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930076 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930097 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930117 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930143 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930185 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930261 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930272 4979 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/88d88d25-879c-461c-a146-89bc738ef3ff-logs\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930281 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqp64\" (UniqueName: \"kubernetes.io/projected/88d88d25-879c-461c-a146-89bc738ef3ff-kube-api-access-rqp64\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:12 crc kubenswrapper[4979]: I1204 14:11:12.930289 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88d88d25-879c-461c-a146-89bc738ef3ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.031688 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.031793 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.031820 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.031851 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.031892 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.031948 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.032077 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpbfq\" (UniqueName: \"kubernetes.io/projected/351d7f45-a24f-46da-a9c9-2b598fff069d-kube-api-access-lpbfq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.032132 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.032165 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.032212 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.032231 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.037240 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.045055 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.052042 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.052091 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.062842 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-1\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.068053 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-0\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.071509 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ceph\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.074057 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ssh-key\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.082843 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-inventory\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.089242 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-combined-ca-bundle\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.111005 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpbfq\" (UniqueName: \"kubernetes.io/projected/351d7f45-a24f-46da-a9c9-2b598fff069d-kube-api-access-lpbfq\") pod \"nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.165837 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.274770 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"88d88d25-879c-461c-a146-89bc738ef3ff","Type":"ContainerDied","Data":"1a14811c51094cd3380d0ab77fb3b5596e1c8f5517083ea416c6ee40c268d4bd"} Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.275024 4979 scope.go:117] "RemoveContainer" containerID="ce04a68a8d5882a2a8474677b476c012dad476be2cac35c4de03abbf3a0a803b" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.275176 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.301538 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"bc6d010c-f547-4ea1-9d88-16a8a85834bf","Type":"ContainerDied","Data":"fcb5adece712c264be7f9a63d7c7f25cbc9f667c3cdf8e909c277b254831140d"} Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.301634 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.354162 4979 scope.go:117] "RemoveContainer" containerID="e15b275173b8bd8e14ab47413fbf97633b85db28701be0e8cbdb009087d43c9f" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.364428 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.391557 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.405003 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.417402 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.425247 4979 scope.go:117] "RemoveContainer" containerID="d0a361d3f246a9a0b46fcb15732158eba4b629166f2aa067d3ab69b289653692" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.431459 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.434519 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.441536 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.443778 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.458712 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.461221 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.466549 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.489669 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.554919 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6124ed0-efd3-4275-982f-994549126581-config-data\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.555018 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vzwb\" (UniqueName: \"kubernetes.io/projected/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-kube-api-access-8vzwb\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.555882 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-logs\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.556043 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjqmm\" (UniqueName: \"kubernetes.io/projected/c6124ed0-efd3-4275-982f-994549126581-kube-api-access-kjqmm\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.556253 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6124ed0-efd3-4275-982f-994549126581-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.556348 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.556503 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-config-data\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.557066 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6124ed0-efd3-4275-982f-994549126581-logs\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.568472 4979 scope.go:117] "RemoveContainer" containerID="9f1eda5655293e5c52af5fb24746e23eec2651f8e849d13062d637d2cf6d7f81" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658594 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6124ed0-efd3-4275-982f-994549126581-logs\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658690 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6124ed0-efd3-4275-982f-994549126581-config-data\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658735 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vzwb\" (UniqueName: \"kubernetes.io/projected/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-kube-api-access-8vzwb\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658764 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-logs\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658817 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjqmm\" (UniqueName: \"kubernetes.io/projected/c6124ed0-efd3-4275-982f-994549126581-kube-api-access-kjqmm\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658891 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6124ed0-efd3-4275-982f-994549126581-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658924 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.658951 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-config-data\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.659951 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c6124ed0-efd3-4275-982f-994549126581-logs\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.661806 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-logs\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.665533 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.668761 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6124ed0-efd3-4275-982f-994549126581-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.669360 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6124ed0-efd3-4275-982f-994549126581-config-data\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.669637 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-config-data\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.677910 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjqmm\" (UniqueName: \"kubernetes.io/projected/c6124ed0-efd3-4275-982f-994549126581-kube-api-access-kjqmm\") pod \"nova-api-0\" (UID: \"c6124ed0-efd3-4275-982f-994549126581\") " pod="openstack/nova-api-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.680943 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vzwb\" (UniqueName: \"kubernetes.io/projected/b3cd38ce-827c-43bd-84ad-c676c62c8eaf-kube-api-access-8vzwb\") pod \"nova-metadata-0\" (UID: \"b3cd38ce-827c-43bd-84ad-c676c62c8eaf\") " pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.804878 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 04 14:11:13 crc kubenswrapper[4979]: I1204 14:11:13.809922 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.040801 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts"] Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.050322 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.216464 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88d88d25-879c-461c-a146-89bc738ef3ff" path="/var/lib/kubelet/pods/88d88d25-879c-461c-a146-89bc738ef3ff/volumes" Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.217856 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc6d010c-f547-4ea1-9d88-16a8a85834bf" path="/var/lib/kubelet/pods/bc6d010c-f547-4ea1-9d88-16a8a85834bf/volumes" Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.314584 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" event={"ID":"351d7f45-a24f-46da-a9c9-2b598fff069d","Type":"ContainerStarted","Data":"b674e6ad8288a3e296d4709aff275b1630a2fce6ec4da2cc3be228d440ac0ca9"} Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.396967 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 04 14:11:14 crc kubenswrapper[4979]: I1204 14:11:14.469499 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 04 14:11:14 crc kubenswrapper[4979]: W1204 14:11:14.885665 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc6124ed0_efd3_4275_982f_994549126581.slice/crio-934b7f05e650e33cb8a9e872a40660a17d968205c7d81717141af899b47f7298 WatchSource:0}: Error finding container 934b7f05e650e33cb8a9e872a40660a17d968205c7d81717141af899b47f7298: Status 404 returned error can't find the container with id 934b7f05e650e33cb8a9e872a40660a17d968205c7d81717141af899b47f7298 Dec 04 14:11:15 crc kubenswrapper[4979]: I1204 14:11:15.336178 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cd38ce-827c-43bd-84ad-c676c62c8eaf","Type":"ContainerStarted","Data":"97e924e62777381846cf3731d406e1211c15bdc787a3ac9bca8c49622834390c"} Dec 04 14:11:15 crc kubenswrapper[4979]: I1204 14:11:15.338072 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c6124ed0-efd3-4275-982f-994549126581","Type":"ContainerStarted","Data":"934b7f05e650e33cb8a9e872a40660a17d968205c7d81717141af899b47f7298"} Dec 04 14:11:15 crc kubenswrapper[4979]: E1204 14:11:15.835400 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87 is running failed: container process not found" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:11:15 crc kubenswrapper[4979]: E1204 14:11:15.836580 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87 is running failed: container process not found" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:11:15 crc kubenswrapper[4979]: E1204 14:11:15.837430 4979 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87 is running failed: container process not found" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 04 14:11:15 crc kubenswrapper[4979]: E1204 14:11:15.837487 4979 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="cf1145a5-ea36-490c-9360-98feed19a052" containerName="nova-scheduler-scheduler" Dec 04 14:11:15 crc kubenswrapper[4979]: I1204 14:11:15.979606 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.122056 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-config-data\") pod \"cf1145a5-ea36-490c-9360-98feed19a052\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.122151 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-combined-ca-bundle\") pod \"cf1145a5-ea36-490c-9360-98feed19a052\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.122339 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqzq8\" (UniqueName: \"kubernetes.io/projected/cf1145a5-ea36-490c-9360-98feed19a052-kube-api-access-zqzq8\") pod \"cf1145a5-ea36-490c-9360-98feed19a052\" (UID: \"cf1145a5-ea36-490c-9360-98feed19a052\") " Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.131565 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf1145a5-ea36-490c-9360-98feed19a052-kube-api-access-zqzq8" (OuterVolumeSpecName: "kube-api-access-zqzq8") pod "cf1145a5-ea36-490c-9360-98feed19a052" (UID: "cf1145a5-ea36-490c-9360-98feed19a052"). InnerVolumeSpecName "kube-api-access-zqzq8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.159447 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf1145a5-ea36-490c-9360-98feed19a052" (UID: "cf1145a5-ea36-490c-9360-98feed19a052"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.164642 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-config-data" (OuterVolumeSpecName: "config-data") pod "cf1145a5-ea36-490c-9360-98feed19a052" (UID: "cf1145a5-ea36-490c-9360-98feed19a052"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.224317 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqzq8\" (UniqueName: \"kubernetes.io/projected/cf1145a5-ea36-490c-9360-98feed19a052-kube-api-access-zqzq8\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.224553 4979 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-config-data\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.224640 4979 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf1145a5-ea36-490c-9360-98feed19a052-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.349780 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" event={"ID":"351d7f45-a24f-46da-a9c9-2b598fff069d","Type":"ContainerStarted","Data":"d7443e190d7e4de5c0b3b71535726d91fea5817eb474884427a90d71c9806118"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.353452 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cd38ce-827c-43bd-84ad-c676c62c8eaf","Type":"ContainerStarted","Data":"f6ba59618fad0b263b12848e3bde332ed6962cb72e46d47925e990b52d418a7d"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.353520 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b3cd38ce-827c-43bd-84ad-c676c62c8eaf","Type":"ContainerStarted","Data":"b73b84c2823a8bef7d547e853cf0d1dddfe577393ba23b6588b5bed89a2ee8c2"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.355516 4979 generic.go:334] "Generic (PLEG): container finished" podID="cf1145a5-ea36-490c-9360-98feed19a052" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" exitCode=0 Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.355561 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.355580 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cf1145a5-ea36-490c-9360-98feed19a052","Type":"ContainerDied","Data":"642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.356332 4979 scope.go:117] "RemoveContainer" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.356326 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"cf1145a5-ea36-490c-9360-98feed19a052","Type":"ContainerDied","Data":"24049e61222c2c7adbb5589b7c9c01c329d9b4e387a14f345d2ab440076566c5"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.358133 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c6124ed0-efd3-4275-982f-994549126581","Type":"ContainerStarted","Data":"a0bcb7741c2766df7427699682eb4499444e2627cd9d357ec54eaf99c5b5e7e1"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.358160 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"c6124ed0-efd3-4275-982f-994549126581","Type":"ContainerStarted","Data":"6b126686a7fad9946ac80fa2ece0f08f3a1b8492bdaf8caf4e55c3dbe22ffdd8"} Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.384566 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" podStartSLOduration=3.463204104 podStartE2EDuration="4.384545626s" podCreationTimestamp="2025-12-04 14:11:12 +0000 UTC" firstStartedPulling="2025-12-04 14:11:14.050057522 +0000 UTC m=+8898.324353326" lastFinishedPulling="2025-12-04 14:11:14.971399034 +0000 UTC m=+8899.245694848" observedRunningTime="2025-12-04 14:11:16.377717367 +0000 UTC m=+8900.652013201" watchObservedRunningTime="2025-12-04 14:11:16.384545626 +0000 UTC m=+8900.658841430" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.391683 4979 scope.go:117] "RemoveContainer" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" Dec 04 14:11:16 crc kubenswrapper[4979]: E1204 14:11:16.392046 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87\": container with ID starting with 642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87 not found: ID does not exist" containerID="642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.392069 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87"} err="failed to get container status \"642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87\": rpc error: code = NotFound desc = could not find container \"642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87\": container with ID starting with 642cac32fee71c9ea2961e56e50ce46f7e07dd8c102e57e8cdbe99b821ca5a87 not found: ID does not exist" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.409342 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.409321827 podStartE2EDuration="3.409321827s" podCreationTimestamp="2025-12-04 14:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:16.397854916 +0000 UTC m=+8900.672150800" watchObservedRunningTime="2025-12-04 14:11:16.409321827 +0000 UTC m=+8900.683617631" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.433396 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.433377849 podStartE2EDuration="3.433377849s" podCreationTimestamp="2025-12-04 14:11:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:16.423452708 +0000 UTC m=+8900.697748512" watchObservedRunningTime="2025-12-04 14:11:16.433377849 +0000 UTC m=+8900.707673653" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.458731 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.483105 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.502282 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:11:16 crc kubenswrapper[4979]: E1204 14:11:16.524519 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf1145a5-ea36-490c-9360-98feed19a052" containerName="nova-scheduler-scheduler" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.524541 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf1145a5-ea36-490c-9360-98feed19a052" containerName="nova-scheduler-scheduler" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.524924 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf1145a5-ea36-490c-9360-98feed19a052" containerName="nova-scheduler-scheduler" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.527012 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.528670 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.533619 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.639814 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.639899 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v54mb\" (UniqueName: \"kubernetes.io/projected/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-kube-api-access-v54mb\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.640041 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-config-data\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.742223 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.742292 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v54mb\" (UniqueName: \"kubernetes.io/projected/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-kube-api-access-v54mb\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.742410 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-config-data\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.972874 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v54mb\" (UniqueName: \"kubernetes.io/projected/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-kube-api-access-v54mb\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.972998 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-config-data\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:16 crc kubenswrapper[4979]: I1204 14:11:16.975061 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55b8d9b5-db3e-41fa-b712-02f19a7f3bfe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe\") " pod="openstack/nova-scheduler-0" Dec 04 14:11:17 crc kubenswrapper[4979]: I1204 14:11:17.150521 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 04 14:11:17 crc kubenswrapper[4979]: I1204 14:11:17.635464 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 04 14:11:17 crc kubenswrapper[4979]: W1204 14:11:17.638558 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod55b8d9b5_db3e_41fa_b712_02f19a7f3bfe.slice/crio-82b96db022539d234d21d63577e290a31730564a77717f9f5a25276c95fb628a WatchSource:0}: Error finding container 82b96db022539d234d21d63577e290a31730564a77717f9f5a25276c95fb628a: Status 404 returned error can't find the container with id 82b96db022539d234d21d63577e290a31730564a77717f9f5a25276c95fb628a Dec 04 14:11:18 crc kubenswrapper[4979]: I1204 14:11:18.224419 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf1145a5-ea36-490c-9360-98feed19a052" path="/var/lib/kubelet/pods/cf1145a5-ea36-490c-9360-98feed19a052/volumes" Dec 04 14:11:18 crc kubenswrapper[4979]: I1204 14:11:18.379833 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe","Type":"ContainerStarted","Data":"1afac260b8c2d65c4eb176b4ef3c4182a6f3ad4d74e2252d8c885f19652753c0"} Dec 04 14:11:18 crc kubenswrapper[4979]: I1204 14:11:18.380080 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"55b8d9b5-db3e-41fa-b712-02f19a7f3bfe","Type":"ContainerStarted","Data":"82b96db022539d234d21d63577e290a31730564a77717f9f5a25276c95fb628a"} Dec 04 14:11:18 crc kubenswrapper[4979]: I1204 14:11:18.409733 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.409714816 podStartE2EDuration="2.409714816s" podCreationTimestamp="2025-12-04 14:11:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-04 14:11:18.403674437 +0000 UTC m=+8902.677970251" watchObservedRunningTime="2025-12-04 14:11:18.409714816 +0000 UTC m=+8902.684010620" Dec 04 14:11:18 crc kubenswrapper[4979]: I1204 14:11:18.805251 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:11:18 crc kubenswrapper[4979]: I1204 14:11:18.805737 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 04 14:11:19 crc kubenswrapper[4979]: I1204 14:11:19.652310 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 04 14:11:19 crc kubenswrapper[4979]: I1204 14:11:19.653285 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 04 14:11:22 crc kubenswrapper[4979]: I1204 14:11:22.151159 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 04 14:11:23 crc kubenswrapper[4979]: I1204 14:11:23.805343 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:11:23 crc kubenswrapper[4979]: I1204 14:11:23.805669 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 04 14:11:23 crc kubenswrapper[4979]: I1204 14:11:23.811473 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:11:23 crc kubenswrapper[4979]: I1204 14:11:23.811513 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 04 14:11:24 crc kubenswrapper[4979]: I1204 14:11:24.887612 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b3cd38ce-827c-43bd-84ad-c676c62c8eaf" containerName="nova-metadata-log" probeResult="failure" output="Get \"http://10.217.1.195:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:11:24 crc kubenswrapper[4979]: I1204 14:11:24.969726 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c6124ed0-efd3-4275-982f-994549126581" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:11:24 crc kubenswrapper[4979]: I1204 14:11:24.978033 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="c6124ed0-efd3-4275-982f-994549126581" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.194:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:11:24 crc kubenswrapper[4979]: I1204 14:11:24.981065 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b3cd38ce-827c-43bd-84ad-c676c62c8eaf" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"http://10.217.1.195:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:11:27 crc kubenswrapper[4979]: I1204 14:11:27.151269 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 04 14:11:27 crc kubenswrapper[4979]: I1204 14:11:27.321801 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 04 14:11:27 crc kubenswrapper[4979]: I1204 14:11:27.510393 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 04 14:11:28 crc kubenswrapper[4979]: I1204 14:11:28.041166 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:11:28 crc kubenswrapper[4979]: I1204 14:11:28.041421 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.812909 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.820226 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.820913 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.821371 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.823851 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.824154 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:11:33 crc kubenswrapper[4979]: I1204 14:11:33.831190 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:11:34 crc kubenswrapper[4979]: I1204 14:11:34.563495 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 04 14:11:34 crc kubenswrapper[4979]: I1204 14:11:34.566799 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 04 14:11:34 crc kubenswrapper[4979]: I1204 14:11:34.568253 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 04 14:11:58 crc kubenswrapper[4979]: I1204 14:11:58.040756 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:11:58 crc kubenswrapper[4979]: I1204 14:11:58.041331 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:12:28 crc kubenswrapper[4979]: I1204 14:12:28.040809 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:12:28 crc kubenswrapper[4979]: I1204 14:12:28.042793 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:12:28 crc kubenswrapper[4979]: I1204 14:12:28.043014 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 14:12:28 crc kubenswrapper[4979]: I1204 14:12:28.044116 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2dda8f444bccd60d9a9921986f15443592d20a8839d0aa1af7274bb26ac30114"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:12:28 crc kubenswrapper[4979]: I1204 14:12:28.044372 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://2dda8f444bccd60d9a9921986f15443592d20a8839d0aa1af7274bb26ac30114" gracePeriod=600 Dec 04 14:12:29 crc kubenswrapper[4979]: I1204 14:12:29.123440 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="2dda8f444bccd60d9a9921986f15443592d20a8839d0aa1af7274bb26ac30114" exitCode=0 Dec 04 14:12:29 crc kubenswrapper[4979]: I1204 14:12:29.123518 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"2dda8f444bccd60d9a9921986f15443592d20a8839d0aa1af7274bb26ac30114"} Dec 04 14:12:29 crc kubenswrapper[4979]: I1204 14:12:29.123837 4979 scope.go:117] "RemoveContainer" containerID="445ac5b270aed9350a3eec4201543b4006d6d828047dc4d6fe0ec37d51ec7982" Dec 04 14:12:31 crc kubenswrapper[4979]: I1204 14:12:31.151922 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1"} Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.117201 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tdftb"] Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.122495 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.135890 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tdftb"] Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.289024 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2sjm\" (UniqueName: \"kubernetes.io/projected/91af5405-002e-461d-94a6-e73abb8c6a55-kube-api-access-k2sjm\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.289600 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-utilities\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.289644 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-catalog-content\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.392547 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-utilities\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.392714 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-catalog-content\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.393105 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-utilities\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.393825 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-catalog-content\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.393883 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2sjm\" (UniqueName: \"kubernetes.io/projected/91af5405-002e-461d-94a6-e73abb8c6a55-kube-api-access-k2sjm\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.426607 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2sjm\" (UniqueName: \"kubernetes.io/projected/91af5405-002e-461d-94a6-e73abb8c6a55-kube-api-access-k2sjm\") pod \"redhat-operators-tdftb\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.455055 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:35 crc kubenswrapper[4979]: I1204 14:13:35.934535 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tdftb"] Dec 04 14:13:36 crc kubenswrapper[4979]: I1204 14:13:36.846685 4979 generic.go:334] "Generic (PLEG): container finished" podID="91af5405-002e-461d-94a6-e73abb8c6a55" containerID="f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0" exitCode=0 Dec 04 14:13:36 crc kubenswrapper[4979]: I1204 14:13:36.846744 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerDied","Data":"f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0"} Dec 04 14:13:36 crc kubenswrapper[4979]: I1204 14:13:36.846937 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerStarted","Data":"f1a2c59af94da3fe47afaea3164599f384f7ca37b8c82fd433ec2a6782281236"} Dec 04 14:13:37 crc kubenswrapper[4979]: I1204 14:13:37.857571 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerStarted","Data":"14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513"} Dec 04 14:13:40 crc kubenswrapper[4979]: I1204 14:13:40.891980 4979 generic.go:334] "Generic (PLEG): container finished" podID="91af5405-002e-461d-94a6-e73abb8c6a55" containerID="14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513" exitCode=0 Dec 04 14:13:40 crc kubenswrapper[4979]: I1204 14:13:40.892069 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerDied","Data":"14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513"} Dec 04 14:13:41 crc kubenswrapper[4979]: I1204 14:13:41.906237 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerStarted","Data":"91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c"} Dec 04 14:13:41 crc kubenswrapper[4979]: I1204 14:13:41.934738 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tdftb" podStartSLOduration=2.281973463 podStartE2EDuration="6.934716117s" podCreationTimestamp="2025-12-04 14:13:35 +0000 UTC" firstStartedPulling="2025-12-04 14:13:36.849217085 +0000 UTC m=+9041.123512889" lastFinishedPulling="2025-12-04 14:13:41.501959749 +0000 UTC m=+9045.776255543" observedRunningTime="2025-12-04 14:13:41.928315958 +0000 UTC m=+9046.202611772" watchObservedRunningTime="2025-12-04 14:13:41.934716117 +0000 UTC m=+9046.209011931" Dec 04 14:13:45 crc kubenswrapper[4979]: I1204 14:13:45.455673 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:45 crc kubenswrapper[4979]: I1204 14:13:45.455979 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:46 crc kubenswrapper[4979]: I1204 14:13:46.508839 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tdftb" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="registry-server" probeResult="failure" output=< Dec 04 14:13:46 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 14:13:46 crc kubenswrapper[4979]: > Dec 04 14:13:55 crc kubenswrapper[4979]: I1204 14:13:55.524279 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:55 crc kubenswrapper[4979]: I1204 14:13:55.573030 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:55 crc kubenswrapper[4979]: I1204 14:13:55.765647 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tdftb"] Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.059142 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tdftb" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="registry-server" containerID="cri-o://91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c" gracePeriod=2 Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.643369 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.727025 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-utilities\") pod \"91af5405-002e-461d-94a6-e73abb8c6a55\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.727509 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-catalog-content\") pod \"91af5405-002e-461d-94a6-e73abb8c6a55\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.727700 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2sjm\" (UniqueName: \"kubernetes.io/projected/91af5405-002e-461d-94a6-e73abb8c6a55-kube-api-access-k2sjm\") pod \"91af5405-002e-461d-94a6-e73abb8c6a55\" (UID: \"91af5405-002e-461d-94a6-e73abb8c6a55\") " Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.728241 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-utilities" (OuterVolumeSpecName: "utilities") pod "91af5405-002e-461d-94a6-e73abb8c6a55" (UID: "91af5405-002e-461d-94a6-e73abb8c6a55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.728695 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.734410 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91af5405-002e-461d-94a6-e73abb8c6a55-kube-api-access-k2sjm" (OuterVolumeSpecName: "kube-api-access-k2sjm") pod "91af5405-002e-461d-94a6-e73abb8c6a55" (UID: "91af5405-002e-461d-94a6-e73abb8c6a55"). InnerVolumeSpecName "kube-api-access-k2sjm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.830227 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2sjm\" (UniqueName: \"kubernetes.io/projected/91af5405-002e-461d-94a6-e73abb8c6a55-kube-api-access-k2sjm\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.849619 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91af5405-002e-461d-94a6-e73abb8c6a55" (UID: "91af5405-002e-461d-94a6-e73abb8c6a55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:13:57 crc kubenswrapper[4979]: I1204 14:13:57.932026 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91af5405-002e-461d-94a6-e73abb8c6a55-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.076582 4979 generic.go:334] "Generic (PLEG): container finished" podID="91af5405-002e-461d-94a6-e73abb8c6a55" containerID="91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c" exitCode=0 Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.076687 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerDied","Data":"91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c"} Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.076659 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tdftb" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.076777 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tdftb" event={"ID":"91af5405-002e-461d-94a6-e73abb8c6a55","Type":"ContainerDied","Data":"f1a2c59af94da3fe47afaea3164599f384f7ca37b8c82fd433ec2a6782281236"} Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.076806 4979 scope.go:117] "RemoveContainer" containerID="91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.115833 4979 scope.go:117] "RemoveContainer" containerID="14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.127502 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tdftb"] Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.141598 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tdftb"] Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.176451 4979 scope.go:117] "RemoveContainer" containerID="f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.210888 4979 scope.go:117] "RemoveContainer" containerID="91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c" Dec 04 14:13:58 crc kubenswrapper[4979]: E1204 14:13:58.211342 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c\": container with ID starting with 91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c not found: ID does not exist" containerID="91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.211383 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c"} err="failed to get container status \"91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c\": rpc error: code = NotFound desc = could not find container \"91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c\": container with ID starting with 91b3a18c0bff94613a9c2cf9b1c901480d3978c9e4e5d702f776f2243b292d9c not found: ID does not exist" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.211406 4979 scope.go:117] "RemoveContainer" containerID="14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513" Dec 04 14:13:58 crc kubenswrapper[4979]: E1204 14:13:58.211756 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513\": container with ID starting with 14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513 not found: ID does not exist" containerID="14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.211787 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513"} err="failed to get container status \"14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513\": rpc error: code = NotFound desc = could not find container \"14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513\": container with ID starting with 14a30f57e63db8f54841189dcb63278424c9289c538703b1e009c70ec8065513 not found: ID does not exist" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.211808 4979 scope.go:117] "RemoveContainer" containerID="f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0" Dec 04 14:13:58 crc kubenswrapper[4979]: E1204 14:13:58.212124 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0\": container with ID starting with f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0 not found: ID does not exist" containerID="f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.212170 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0"} err="failed to get container status \"f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0\": rpc error: code = NotFound desc = could not find container \"f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0\": container with ID starting with f9f0ebd42f1a9d114f39daaa53a43f0c565b7530eaa214b5f2c12ffa4d9e9fd0 not found: ID does not exist" Dec 04 14:13:58 crc kubenswrapper[4979]: I1204 14:13:58.218818 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" path="/var/lib/kubelet/pods/91af5405-002e-461d-94a6-e73abb8c6a55/volumes" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.902695 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-525m4"] Dec 04 14:14:38 crc kubenswrapper[4979]: E1204 14:14:38.906898 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="extract-content" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.906925 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="extract-content" Dec 04 14:14:38 crc kubenswrapper[4979]: E1204 14:14:38.906948 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="extract-utilities" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.906959 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="extract-utilities" Dec 04 14:14:38 crc kubenswrapper[4979]: E1204 14:14:38.907012 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="registry-server" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.907021 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="registry-server" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.907414 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="91af5405-002e-461d-94a6-e73abb8c6a55" containerName="registry-server" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.909466 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:38 crc kubenswrapper[4979]: I1204 14:14:38.926615 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-525m4"] Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.025364 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-catalog-content\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.025432 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-utilities\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.025501 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n87mv\" (UniqueName: \"kubernetes.io/projected/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-kube-api-access-n87mv\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.127914 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-catalog-content\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.127986 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-utilities\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.128041 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n87mv\" (UniqueName: \"kubernetes.io/projected/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-kube-api-access-n87mv\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.128462 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-catalog-content\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.128472 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-utilities\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.147270 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n87mv\" (UniqueName: \"kubernetes.io/projected/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-kube-api-access-n87mv\") pod \"community-operators-525m4\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.247031 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:39 crc kubenswrapper[4979]: I1204 14:14:39.819644 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-525m4"] Dec 04 14:14:39 crc kubenswrapper[4979]: W1204 14:14:39.825713 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5300b0b3_57f1_448d_968e_0ba96aa1bd3e.slice/crio-a55d89d387ab6d5a53551c583f3527c93bb387c5fcec56b4a3a6a65aebb8f188 WatchSource:0}: Error finding container a55d89d387ab6d5a53551c583f3527c93bb387c5fcec56b4a3a6a65aebb8f188: Status 404 returned error can't find the container with id a55d89d387ab6d5a53551c583f3527c93bb387c5fcec56b4a3a6a65aebb8f188 Dec 04 14:14:40 crc kubenswrapper[4979]: I1204 14:14:40.548671 4979 generic.go:334] "Generic (PLEG): container finished" podID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerID="7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d" exitCode=0 Dec 04 14:14:40 crc kubenswrapper[4979]: I1204 14:14:40.548860 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerDied","Data":"7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d"} Dec 04 14:14:40 crc kubenswrapper[4979]: I1204 14:14:40.549010 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerStarted","Data":"a55d89d387ab6d5a53551c583f3527c93bb387c5fcec56b4a3a6a65aebb8f188"} Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.498256 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-62f22"] Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.501286 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.517730 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62f22"] Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.567597 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerStarted","Data":"870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3"} Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.590080 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p29tj\" (UniqueName: \"kubernetes.io/projected/2e579ae4-5d38-4f77-a19a-c540f2bbea39-kube-api-access-p29tj\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.590320 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-utilities\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.590818 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-catalog-content\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.693447 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-catalog-content\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.693621 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p29tj\" (UniqueName: \"kubernetes.io/projected/2e579ae4-5d38-4f77-a19a-c540f2bbea39-kube-api-access-p29tj\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.693704 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-utilities\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.693805 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-catalog-content\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.694101 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-utilities\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.712750 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p29tj\" (UniqueName: \"kubernetes.io/projected/2e579ae4-5d38-4f77-a19a-c540f2bbea39-kube-api-access-p29tj\") pod \"certified-operators-62f22\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:41 crc kubenswrapper[4979]: I1204 14:14:41.827109 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:42 crc kubenswrapper[4979]: I1204 14:14:42.435871 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62f22"] Dec 04 14:14:42 crc kubenswrapper[4979]: W1204 14:14:42.437637 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e579ae4_5d38_4f77_a19a_c540f2bbea39.slice/crio-6007a82b06e41736b1f73991e4f467015116dc573a11a8caa77b144a525f7168 WatchSource:0}: Error finding container 6007a82b06e41736b1f73991e4f467015116dc573a11a8caa77b144a525f7168: Status 404 returned error can't find the container with id 6007a82b06e41736b1f73991e4f467015116dc573a11a8caa77b144a525f7168 Dec 04 14:14:42 crc kubenswrapper[4979]: I1204 14:14:42.577835 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerStarted","Data":"6007a82b06e41736b1f73991e4f467015116dc573a11a8caa77b144a525f7168"} Dec 04 14:14:42 crc kubenswrapper[4979]: I1204 14:14:42.580199 4979 generic.go:334] "Generic (PLEG): container finished" podID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerID="870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3" exitCode=0 Dec 04 14:14:42 crc kubenswrapper[4979]: I1204 14:14:42.580256 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerDied","Data":"870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3"} Dec 04 14:14:43 crc kubenswrapper[4979]: I1204 14:14:43.591480 4979 generic.go:334] "Generic (PLEG): container finished" podID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerID="1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d" exitCode=0 Dec 04 14:14:43 crc kubenswrapper[4979]: I1204 14:14:43.591774 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerDied","Data":"1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d"} Dec 04 14:14:44 crc kubenswrapper[4979]: I1204 14:14:44.604251 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerStarted","Data":"484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162"} Dec 04 14:14:44 crc kubenswrapper[4979]: I1204 14:14:44.632535 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-525m4" podStartSLOduration=4.058249608 podStartE2EDuration="6.632514073s" podCreationTimestamp="2025-12-04 14:14:38 +0000 UTC" firstStartedPulling="2025-12-04 14:14:40.554022953 +0000 UTC m=+9104.828318757" lastFinishedPulling="2025-12-04 14:14:43.128287418 +0000 UTC m=+9107.402583222" observedRunningTime="2025-12-04 14:14:44.624603155 +0000 UTC m=+9108.898898949" watchObservedRunningTime="2025-12-04 14:14:44.632514073 +0000 UTC m=+9108.906809877" Dec 04 14:14:45 crc kubenswrapper[4979]: I1204 14:14:45.616506 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerStarted","Data":"1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97"} Dec 04 14:14:47 crc kubenswrapper[4979]: I1204 14:14:47.640036 4979 generic.go:334] "Generic (PLEG): container finished" podID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerID="1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97" exitCode=0 Dec 04 14:14:47 crc kubenswrapper[4979]: I1204 14:14:47.640082 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerDied","Data":"1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97"} Dec 04 14:14:49 crc kubenswrapper[4979]: I1204 14:14:49.248083 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:49 crc kubenswrapper[4979]: I1204 14:14:49.248736 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:49 crc kubenswrapper[4979]: I1204 14:14:49.307145 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:49 crc kubenswrapper[4979]: I1204 14:14:49.659413 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerStarted","Data":"74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00"} Dec 04 14:14:49 crc kubenswrapper[4979]: I1204 14:14:49.680076 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-62f22" podStartSLOduration=3.551984887 podStartE2EDuration="8.680055987s" podCreationTimestamp="2025-12-04 14:14:41 +0000 UTC" firstStartedPulling="2025-12-04 14:14:43.593528219 +0000 UTC m=+9107.867824023" lastFinishedPulling="2025-12-04 14:14:48.721599319 +0000 UTC m=+9112.995895123" observedRunningTime="2025-12-04 14:14:49.678050584 +0000 UTC m=+9113.952346388" watchObservedRunningTime="2025-12-04 14:14:49.680055987 +0000 UTC m=+9113.954351791" Dec 04 14:14:49 crc kubenswrapper[4979]: I1204 14:14:49.714539 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:51 crc kubenswrapper[4979]: I1204 14:14:51.082643 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-525m4"] Dec 04 14:14:51 crc kubenswrapper[4979]: I1204 14:14:51.675903 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-525m4" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="registry-server" containerID="cri-o://484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162" gracePeriod=2 Dec 04 14:14:51 crc kubenswrapper[4979]: I1204 14:14:51.827762 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:51 crc kubenswrapper[4979]: I1204 14:14:51.827848 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:51 crc kubenswrapper[4979]: I1204 14:14:51.884688 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.456478 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.565447 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n87mv\" (UniqueName: \"kubernetes.io/projected/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-kube-api-access-n87mv\") pod \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.565671 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-utilities\") pod \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.565816 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-catalog-content\") pod \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\" (UID: \"5300b0b3-57f1-448d-968e-0ba96aa1bd3e\") " Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.575662 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-kube-api-access-n87mv" (OuterVolumeSpecName: "kube-api-access-n87mv") pod "5300b0b3-57f1-448d-968e-0ba96aa1bd3e" (UID: "5300b0b3-57f1-448d-968e-0ba96aa1bd3e"). InnerVolumeSpecName "kube-api-access-n87mv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.581765 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-utilities" (OuterVolumeSpecName: "utilities") pod "5300b0b3-57f1-448d-968e-0ba96aa1bd3e" (UID: "5300b0b3-57f1-448d-968e-0ba96aa1bd3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.620818 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5300b0b3-57f1-448d-968e-0ba96aa1bd3e" (UID: "5300b0b3-57f1-448d-968e-0ba96aa1bd3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.672717 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.672755 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n87mv\" (UniqueName: \"kubernetes.io/projected/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-kube-api-access-n87mv\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.672768 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5300b0b3-57f1-448d-968e-0ba96aa1bd3e-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.702835 4979 generic.go:334] "Generic (PLEG): container finished" podID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerID="484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162" exitCode=0 Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.702890 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerDied","Data":"484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162"} Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.702922 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-525m4" event={"ID":"5300b0b3-57f1-448d-968e-0ba96aa1bd3e","Type":"ContainerDied","Data":"a55d89d387ab6d5a53551c583f3527c93bb387c5fcec56b4a3a6a65aebb8f188"} Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.702942 4979 scope.go:117] "RemoveContainer" containerID="484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.703113 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-525m4" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.736368 4979 scope.go:117] "RemoveContainer" containerID="870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.743913 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-525m4"] Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.760864 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-525m4"] Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.768645 4979 scope.go:117] "RemoveContainer" containerID="7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.833791 4979 scope.go:117] "RemoveContainer" containerID="484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162" Dec 04 14:14:53 crc kubenswrapper[4979]: E1204 14:14:53.834168 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162\": container with ID starting with 484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162 not found: ID does not exist" containerID="484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.834197 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162"} err="failed to get container status \"484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162\": rpc error: code = NotFound desc = could not find container \"484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162\": container with ID starting with 484e4866f6bfc7590b87e96327c5aa2d4b8d94cd0b94cb362c970f8cf218e162 not found: ID does not exist" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.834216 4979 scope.go:117] "RemoveContainer" containerID="870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3" Dec 04 14:14:53 crc kubenswrapper[4979]: E1204 14:14:53.834582 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3\": container with ID starting with 870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3 not found: ID does not exist" containerID="870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.834602 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3"} err="failed to get container status \"870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3\": rpc error: code = NotFound desc = could not find container \"870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3\": container with ID starting with 870226f9608a40a03ce37514a7ae9ea42392f423ff6eabfaaadad246faad6da3 not found: ID does not exist" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.834614 4979 scope.go:117] "RemoveContainer" containerID="7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d" Dec 04 14:14:53 crc kubenswrapper[4979]: E1204 14:14:53.834967 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d\": container with ID starting with 7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d not found: ID does not exist" containerID="7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d" Dec 04 14:14:53 crc kubenswrapper[4979]: I1204 14:14:53.834990 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d"} err="failed to get container status \"7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d\": rpc error: code = NotFound desc = could not find container \"7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d\": container with ID starting with 7906fb4ae56cc557cf3289f4a7b3a91af144df1e93456e7db8c80a80cc263d1d not found: ID does not exist" Dec 04 14:14:54 crc kubenswrapper[4979]: I1204 14:14:54.212758 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" path="/var/lib/kubelet/pods/5300b0b3-57f1-448d-968e-0ba96aa1bd3e/volumes" Dec 04 14:14:58 crc kubenswrapper[4979]: I1204 14:14:58.040384 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:14:58 crc kubenswrapper[4979]: I1204 14:14:58.041082 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.144127 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs"] Dec 04 14:15:00 crc kubenswrapper[4979]: E1204 14:15:00.145154 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="registry-server" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.145174 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="registry-server" Dec 04 14:15:00 crc kubenswrapper[4979]: E1204 14:15:00.145186 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="extract-utilities" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.145194 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="extract-utilities" Dec 04 14:15:00 crc kubenswrapper[4979]: E1204 14:15:00.145251 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="extract-content" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.145260 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="extract-content" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.145867 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="5300b0b3-57f1-448d-968e-0ba96aa1bd3e" containerName="registry-server" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.146730 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.152525 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.152628 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.182846 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs"] Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.321053 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68bd9413-1093-4bdc-b0eb-164835ea9a9b-secret-volume\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.321139 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvxz7\" (UniqueName: \"kubernetes.io/projected/68bd9413-1093-4bdc-b0eb-164835ea9a9b-kube-api-access-cvxz7\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.321288 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68bd9413-1093-4bdc-b0eb-164835ea9a9b-config-volume\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.423387 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68bd9413-1093-4bdc-b0eb-164835ea9a9b-secret-volume\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.423458 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvxz7\" (UniqueName: \"kubernetes.io/projected/68bd9413-1093-4bdc-b0eb-164835ea9a9b-kube-api-access-cvxz7\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.423530 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68bd9413-1093-4bdc-b0eb-164835ea9a9b-config-volume\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.424774 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68bd9413-1093-4bdc-b0eb-164835ea9a9b-config-volume\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.441796 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68bd9413-1093-4bdc-b0eb-164835ea9a9b-secret-volume\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.452136 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvxz7\" (UniqueName: \"kubernetes.io/projected/68bd9413-1093-4bdc-b0eb-164835ea9a9b-kube-api-access-cvxz7\") pod \"collect-profiles-29414295-qbhrs\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:00 crc kubenswrapper[4979]: I1204 14:15:00.524535 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:01 crc kubenswrapper[4979]: I1204 14:15:01.051975 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs"] Dec 04 14:15:01 crc kubenswrapper[4979]: I1204 14:15:01.786864 4979 generic.go:334] "Generic (PLEG): container finished" podID="68bd9413-1093-4bdc-b0eb-164835ea9a9b" containerID="2624ee4600cda03ac1bbc53c86a6857ba3c05f37e389103d2e187a7352ca6b8b" exitCode=0 Dec 04 14:15:01 crc kubenswrapper[4979]: I1204 14:15:01.786978 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" event={"ID":"68bd9413-1093-4bdc-b0eb-164835ea9a9b","Type":"ContainerDied","Data":"2624ee4600cda03ac1bbc53c86a6857ba3c05f37e389103d2e187a7352ca6b8b"} Dec 04 14:15:01 crc kubenswrapper[4979]: I1204 14:15:01.787431 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" event={"ID":"68bd9413-1093-4bdc-b0eb-164835ea9a9b","Type":"ContainerStarted","Data":"8e2b93955aff6891495c87899ecc574f53fa49e748faefb3bda26b888119126f"} Dec 04 14:15:01 crc kubenswrapper[4979]: I1204 14:15:01.885140 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:15:01 crc kubenswrapper[4979]: I1204 14:15:01.940243 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62f22"] Dec 04 14:15:02 crc kubenswrapper[4979]: I1204 14:15:02.800417 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-62f22" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="registry-server" containerID="cri-o://74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00" gracePeriod=2 Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.231897 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.300821 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvxz7\" (UniqueName: \"kubernetes.io/projected/68bd9413-1093-4bdc-b0eb-164835ea9a9b-kube-api-access-cvxz7\") pod \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.300933 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68bd9413-1093-4bdc-b0eb-164835ea9a9b-config-volume\") pod \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.301020 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68bd9413-1093-4bdc-b0eb-164835ea9a9b-secret-volume\") pod \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\" (UID: \"68bd9413-1093-4bdc-b0eb-164835ea9a9b\") " Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.302043 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68bd9413-1093-4bdc-b0eb-164835ea9a9b-config-volume" (OuterVolumeSpecName: "config-volume") pod "68bd9413-1093-4bdc-b0eb-164835ea9a9b" (UID: "68bd9413-1093-4bdc-b0eb-164835ea9a9b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.303379 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/68bd9413-1093-4bdc-b0eb-164835ea9a9b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.308606 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68bd9413-1093-4bdc-b0eb-164835ea9a9b-kube-api-access-cvxz7" (OuterVolumeSpecName: "kube-api-access-cvxz7") pod "68bd9413-1093-4bdc-b0eb-164835ea9a9b" (UID: "68bd9413-1093-4bdc-b0eb-164835ea9a9b"). InnerVolumeSpecName "kube-api-access-cvxz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.324465 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68bd9413-1093-4bdc-b0eb-164835ea9a9b-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "68bd9413-1093-4bdc-b0eb-164835ea9a9b" (UID: "68bd9413-1093-4bdc-b0eb-164835ea9a9b"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.405781 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvxz7\" (UniqueName: \"kubernetes.io/projected/68bd9413-1093-4bdc-b0eb-164835ea9a9b-kube-api-access-cvxz7\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.406110 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/68bd9413-1093-4bdc-b0eb-164835ea9a9b-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.428872 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.507563 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p29tj\" (UniqueName: \"kubernetes.io/projected/2e579ae4-5d38-4f77-a19a-c540f2bbea39-kube-api-access-p29tj\") pod \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.507737 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-catalog-content\") pod \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.507816 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-utilities\") pod \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\" (UID: \"2e579ae4-5d38-4f77-a19a-c540f2bbea39\") " Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.508691 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-utilities" (OuterVolumeSpecName: "utilities") pod "2e579ae4-5d38-4f77-a19a-c540f2bbea39" (UID: "2e579ae4-5d38-4f77-a19a-c540f2bbea39"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.512709 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e579ae4-5d38-4f77-a19a-c540f2bbea39-kube-api-access-p29tj" (OuterVolumeSpecName: "kube-api-access-p29tj") pod "2e579ae4-5d38-4f77-a19a-c540f2bbea39" (UID: "2e579ae4-5d38-4f77-a19a-c540f2bbea39"). InnerVolumeSpecName "kube-api-access-p29tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.570066 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e579ae4-5d38-4f77-a19a-c540f2bbea39" (UID: "2e579ae4-5d38-4f77-a19a-c540f2bbea39"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.610102 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.610162 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p29tj\" (UniqueName: \"kubernetes.io/projected/2e579ae4-5d38-4f77-a19a-c540f2bbea39-kube-api-access-p29tj\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.610179 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e579ae4-5d38-4f77-a19a-c540f2bbea39-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.817937 4979 generic.go:334] "Generic (PLEG): container finished" podID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerID="74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00" exitCode=0 Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.818021 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerDied","Data":"74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00"} Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.818069 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62f22" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.818094 4979 scope.go:117] "RemoveContainer" containerID="74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.818080 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62f22" event={"ID":"2e579ae4-5d38-4f77-a19a-c540f2bbea39","Type":"ContainerDied","Data":"6007a82b06e41736b1f73991e4f467015116dc573a11a8caa77b144a525f7168"} Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.820989 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" event={"ID":"68bd9413-1093-4bdc-b0eb-164835ea9a9b","Type":"ContainerDied","Data":"8e2b93955aff6891495c87899ecc574f53fa49e748faefb3bda26b888119126f"} Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.821027 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414295-qbhrs" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.821047 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e2b93955aff6891495c87899ecc574f53fa49e748faefb3bda26b888119126f" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.850557 4979 scope.go:117] "RemoveContainer" containerID="1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.860226 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-62f22"] Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.890357 4979 scope.go:117] "RemoveContainer" containerID="1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.898910 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-62f22"] Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.939555 4979 scope.go:117] "RemoveContainer" containerID="74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00" Dec 04 14:15:03 crc kubenswrapper[4979]: E1204 14:15:03.940182 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00\": container with ID starting with 74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00 not found: ID does not exist" containerID="74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.940213 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00"} err="failed to get container status \"74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00\": rpc error: code = NotFound desc = could not find container \"74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00\": container with ID starting with 74e9c850906cbcd159e1b36a125bdf95ae2a380fc633ce4eeb20268bb79a4e00 not found: ID does not exist" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.940236 4979 scope.go:117] "RemoveContainer" containerID="1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97" Dec 04 14:15:03 crc kubenswrapper[4979]: E1204 14:15:03.941409 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97\": container with ID starting with 1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97 not found: ID does not exist" containerID="1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.941434 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97"} err="failed to get container status \"1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97\": rpc error: code = NotFound desc = could not find container \"1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97\": container with ID starting with 1c09f5c86835d52abae122e1c602e3c7c0131b74568b082171e1d17e9f641a97 not found: ID does not exist" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.941459 4979 scope.go:117] "RemoveContainer" containerID="1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d" Dec 04 14:15:03 crc kubenswrapper[4979]: E1204 14:15:03.942591 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d\": container with ID starting with 1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d not found: ID does not exist" containerID="1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d" Dec 04 14:15:03 crc kubenswrapper[4979]: I1204 14:15:03.942630 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d"} err="failed to get container status \"1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d\": rpc error: code = NotFound desc = could not find container \"1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d\": container with ID starting with 1e8b40b57d0038ab190545fbd623a57dc4a472f8f78b00dfec4a57532d7d7f4d not found: ID does not exist" Dec 04 14:15:04 crc kubenswrapper[4979]: I1204 14:15:04.220462 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" path="/var/lib/kubelet/pods/2e579ae4-5d38-4f77-a19a-c540f2bbea39/volumes" Dec 04 14:15:04 crc kubenswrapper[4979]: I1204 14:15:04.328578 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl"] Dec 04 14:15:04 crc kubenswrapper[4979]: I1204 14:15:04.342698 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414250-t96pl"] Dec 04 14:15:06 crc kubenswrapper[4979]: I1204 14:15:06.215087 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d314ef80-82e1-4a37-9e06-340d1ca94e89" path="/var/lib/kubelet/pods/d314ef80-82e1-4a37-9e06-340d1ca94e89/volumes" Dec 04 14:15:15 crc kubenswrapper[4979]: I1204 14:15:15.480191 4979 scope.go:117] "RemoveContainer" containerID="d7a4d12effba1fd70f81e3515f84aa4a67bbc5a1e74fe0bcda4e2ae9977e0752" Dec 04 14:15:28 crc kubenswrapper[4979]: I1204 14:15:28.041224 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:15:28 crc kubenswrapper[4979]: I1204 14:15:28.041779 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:15:43 crc kubenswrapper[4979]: I1204 14:15:43.983413 4979 trace.go:236] Trace[2137326815]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-cm5bl" (04-Dec-2025 14:15:42.942) (total time: 1040ms): Dec 04 14:15:43 crc kubenswrapper[4979]: Trace[2137326815]: [1.040872964s] [1.040872964s] END Dec 04 14:15:58 crc kubenswrapper[4979]: I1204 14:15:58.040787 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:15:58 crc kubenswrapper[4979]: I1204 14:15:58.041401 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:15:58 crc kubenswrapper[4979]: I1204 14:15:58.041448 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 14:15:58 crc kubenswrapper[4979]: I1204 14:15:58.042261 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:15:58 crc kubenswrapper[4979]: I1204 14:15:58.042332 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" gracePeriod=600 Dec 04 14:16:00 crc kubenswrapper[4979]: I1204 14:16:00.463162 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" exitCode=0 Dec 04 14:16:00 crc kubenswrapper[4979]: I1204 14:16:00.463250 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1"} Dec 04 14:16:00 crc kubenswrapper[4979]: I1204 14:16:00.463986 4979 scope.go:117] "RemoveContainer" containerID="2dda8f444bccd60d9a9921986f15443592d20a8839d0aa1af7274bb26ac30114" Dec 04 14:16:01 crc kubenswrapper[4979]: E1204 14:16:01.184186 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:16:01 crc kubenswrapper[4979]: I1204 14:16:01.476713 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:16:01 crc kubenswrapper[4979]: E1204 14:16:01.477466 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:16:13 crc kubenswrapper[4979]: I1204 14:16:13.199051 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:16:13 crc kubenswrapper[4979]: E1204 14:16:13.200174 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:16:15 crc kubenswrapper[4979]: I1204 14:16:15.647842 4979 generic.go:334] "Generic (PLEG): container finished" podID="351d7f45-a24f-46da-a9c9-2b598fff069d" containerID="d7443e190d7e4de5c0b3b71535726d91fea5817eb474884427a90d71c9806118" exitCode=0 Dec 04 14:16:15 crc kubenswrapper[4979]: I1204 14:16:15.647937 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" event={"ID":"351d7f45-a24f-46da-a9c9-2b598fff069d","Type":"ContainerDied","Data":"d7443e190d7e4de5c0b3b71535726d91fea5817eb474884427a90d71c9806118"} Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.673500 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" event={"ID":"351d7f45-a24f-46da-a9c9-2b598fff069d","Type":"ContainerDied","Data":"b674e6ad8288a3e296d4709aff275b1630a2fce6ec4da2cc3be228d440ac0ca9"} Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.673938 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b674e6ad8288a3e296d4709aff275b1630a2fce6ec4da2cc3be228d440ac0ca9" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.721101 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.875841 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-1\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.876439 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpbfq\" (UniqueName: \"kubernetes.io/projected/351d7f45-a24f-46da-a9c9-2b598fff069d-kube-api-access-lpbfq\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.876565 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-0\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.876641 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-1\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.876801 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-1\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.876931 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-0\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.877051 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-0\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.877105 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ceph\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.877286 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-combined-ca-bundle\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.877369 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-inventory\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.877499 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ssh-key\") pod \"351d7f45-a24f-46da-a9c9-2b598fff069d\" (UID: \"351d7f45-a24f-46da-a9c9-2b598fff069d\") " Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.898737 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-combined-ca-bundle" (OuterVolumeSpecName: "nova-cell1-combined-ca-bundle") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-cell1-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.898782 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ceph" (OuterVolumeSpecName: "ceph") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.899781 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/351d7f45-a24f-46da-a9c9-2b598fff069d-kube-api-access-lpbfq" (OuterVolumeSpecName: "kube-api-access-lpbfq") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "kube-api-access-lpbfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.907662 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-inventory" (OuterVolumeSpecName: "inventory") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.915458 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.925390 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-1" (OuterVolumeSpecName: "nova-cells-global-config-1") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-cells-global-config-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.932118 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-0" (OuterVolumeSpecName: "nova-cells-global-config-0") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-cells-global-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.933841 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.936504 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.936849 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.945431 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "351d7f45-a24f-46da-a9c9-2b598fff069d" (UID: "351d7f45-a24f-46da-a9c9-2b598fff069d"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982515 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982555 4979 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-inventory\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982565 4979 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982573 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpbfq\" (UniqueName: \"kubernetes.io/projected/351d7f45-a24f-46da-a9c9-2b598fff069d-kube-api-access-lpbfq\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982583 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982591 4979 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982599 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-1\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982608 4979 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982617 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cells-global-config-0\" (UniqueName: \"kubernetes.io/configmap/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cells-global-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982626 4979 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:17 crc kubenswrapper[4979]: I1204 14:16:17.982633 4979 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/351d7f45-a24f-46da-a9c9-2b598fff069d-ceph\") on node \"crc\" DevicePath \"\"" Dec 04 14:16:18 crc kubenswrapper[4979]: I1204 14:16:18.683986 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts" Dec 04 14:16:24 crc kubenswrapper[4979]: I1204 14:16:24.199260 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:16:24 crc kubenswrapper[4979]: E1204 14:16:24.200101 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:16:39 crc kubenswrapper[4979]: I1204 14:16:39.199266 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:16:39 crc kubenswrapper[4979]: E1204 14:16:39.201669 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:16:52 crc kubenswrapper[4979]: I1204 14:16:52.199946 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:16:52 crc kubenswrapper[4979]: E1204 14:16:52.200835 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:17:05 crc kubenswrapper[4979]: I1204 14:17:05.200099 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:17:05 crc kubenswrapper[4979]: E1204 14:17:05.201112 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:17:18 crc kubenswrapper[4979]: I1204 14:17:18.199485 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:17:18 crc kubenswrapper[4979]: E1204 14:17:18.200370 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:17:32 crc kubenswrapper[4979]: I1204 14:17:32.208056 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:17:32 crc kubenswrapper[4979]: E1204 14:17:32.210030 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.941930 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bp825"] Dec 04 14:17:40 crc kubenswrapper[4979]: E1204 14:17:40.943095 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="351d7f45-a24f-46da-a9c9-2b598fff069d" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943196 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="351d7f45-a24f-46da-a9c9-2b598fff069d" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 04 14:17:40 crc kubenswrapper[4979]: E1204 14:17:40.943239 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="registry-server" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943248 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="registry-server" Dec 04 14:17:40 crc kubenswrapper[4979]: E1204 14:17:40.943272 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="extract-content" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943280 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="extract-content" Dec 04 14:17:40 crc kubenswrapper[4979]: E1204 14:17:40.943340 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="extract-utilities" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943351 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="extract-utilities" Dec 04 14:17:40 crc kubenswrapper[4979]: E1204 14:17:40.943377 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68bd9413-1093-4bdc-b0eb-164835ea9a9b" containerName="collect-profiles" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943386 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="68bd9413-1093-4bdc-b0eb-164835ea9a9b" containerName="collect-profiles" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943652 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="68bd9413-1093-4bdc-b0eb-164835ea9a9b" containerName="collect-profiles" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943669 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e579ae4-5d38-4f77-a19a-c540f2bbea39" containerName="registry-server" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.943680 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="351d7f45-a24f-46da-a9c9-2b598fff069d" containerName="nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.946207 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.961198 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp825"] Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.985508 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-utilities\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.985567 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-catalog-content\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:40 crc kubenswrapper[4979]: I1204 14:17:40.986020 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-552z4\" (UniqueName: \"kubernetes.io/projected/e94ab30a-9354-4a29-80b9-6eb568d27ee9-kube-api-access-552z4\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.089893 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-552z4\" (UniqueName: \"kubernetes.io/projected/e94ab30a-9354-4a29-80b9-6eb568d27ee9-kube-api-access-552z4\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.090129 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-utilities\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.090161 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-catalog-content\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.090944 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-catalog-content\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.091116 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-utilities\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.120533 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-552z4\" (UniqueName: \"kubernetes.io/projected/e94ab30a-9354-4a29-80b9-6eb568d27ee9-kube-api-access-552z4\") pod \"redhat-marketplace-bp825\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.275559 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:41 crc kubenswrapper[4979]: I1204 14:17:41.779400 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp825"] Dec 04 14:17:42 crc kubenswrapper[4979]: I1204 14:17:42.604415 4979 generic.go:334] "Generic (PLEG): container finished" podID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerID="75e247e89dd3597fed99e5640d8026540fdde52a76e1ae2ae962d621b7da392f" exitCode=0 Dec 04 14:17:42 crc kubenswrapper[4979]: I1204 14:17:42.604493 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerDied","Data":"75e247e89dd3597fed99e5640d8026540fdde52a76e1ae2ae962d621b7da392f"} Dec 04 14:17:42 crc kubenswrapper[4979]: I1204 14:17:42.605036 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerStarted","Data":"3d3ced1c0d16ed550292b3c2fc8b895dde761a18c4a797d935b1be857ebaf491"} Dec 04 14:17:42 crc kubenswrapper[4979]: I1204 14:17:42.608439 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:17:44 crc kubenswrapper[4979]: I1204 14:17:44.629237 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerStarted","Data":"6255ebb20e5a32267ce7da4e3927b99bf2180929a008804f7937708d89c5b31c"} Dec 04 14:17:45 crc kubenswrapper[4979]: I1204 14:17:45.199648 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:17:45 crc kubenswrapper[4979]: E1204 14:17:45.200235 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:17:45 crc kubenswrapper[4979]: I1204 14:17:45.644419 4979 generic.go:334] "Generic (PLEG): container finished" podID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerID="6255ebb20e5a32267ce7da4e3927b99bf2180929a008804f7937708d89c5b31c" exitCode=0 Dec 04 14:17:45 crc kubenswrapper[4979]: I1204 14:17:45.644474 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerDied","Data":"6255ebb20e5a32267ce7da4e3927b99bf2180929a008804f7937708d89c5b31c"} Dec 04 14:17:47 crc kubenswrapper[4979]: I1204 14:17:47.671449 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerStarted","Data":"e3e9d767f23e3388a560c253a764e7fc9d6370ca1ea16b62b7376591838e4d94"} Dec 04 14:17:51 crc kubenswrapper[4979]: I1204 14:17:51.275687 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:51 crc kubenswrapper[4979]: I1204 14:17:51.275996 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:51 crc kubenswrapper[4979]: I1204 14:17:51.351019 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:17:51 crc kubenswrapper[4979]: I1204 14:17:51.380660 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bp825" podStartSLOduration=7.329880202 podStartE2EDuration="11.380638806s" podCreationTimestamp="2025-12-04 14:17:40 +0000 UTC" firstStartedPulling="2025-12-04 14:17:42.60800561 +0000 UTC m=+9286.882301414" lastFinishedPulling="2025-12-04 14:17:46.658764214 +0000 UTC m=+9290.933060018" observedRunningTime="2025-12-04 14:17:47.703881208 +0000 UTC m=+9291.978177012" watchObservedRunningTime="2025-12-04 14:17:51.380638806 +0000 UTC m=+9295.654934610" Dec 04 14:17:57 crc kubenswrapper[4979]: I1204 14:17:57.199216 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:17:57 crc kubenswrapper[4979]: E1204 14:17:57.200111 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:18:01 crc kubenswrapper[4979]: I1204 14:18:01.336668 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:18:01 crc kubenswrapper[4979]: I1204 14:18:01.385280 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp825"] Dec 04 14:18:01 crc kubenswrapper[4979]: I1204 14:18:01.836069 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bp825" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="registry-server" containerID="cri-o://e3e9d767f23e3388a560c253a764e7fc9d6370ca1ea16b62b7376591838e4d94" gracePeriod=2 Dec 04 14:18:02 crc kubenswrapper[4979]: I1204 14:18:02.857008 4979 generic.go:334] "Generic (PLEG): container finished" podID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerID="e3e9d767f23e3388a560c253a764e7fc9d6370ca1ea16b62b7376591838e4d94" exitCode=0 Dec 04 14:18:02 crc kubenswrapper[4979]: I1204 14:18:02.857057 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerDied","Data":"e3e9d767f23e3388a560c253a764e7fc9d6370ca1ea16b62b7376591838e4d94"} Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.183459 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.347645 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-utilities\") pod \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.347706 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-catalog-content\") pod \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.347846 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-552z4\" (UniqueName: \"kubernetes.io/projected/e94ab30a-9354-4a29-80b9-6eb568d27ee9-kube-api-access-552z4\") pod \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\" (UID: \"e94ab30a-9354-4a29-80b9-6eb568d27ee9\") " Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.348824 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-utilities" (OuterVolumeSpecName: "utilities") pod "e94ab30a-9354-4a29-80b9-6eb568d27ee9" (UID: "e94ab30a-9354-4a29-80b9-6eb568d27ee9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.356965 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94ab30a-9354-4a29-80b9-6eb568d27ee9-kube-api-access-552z4" (OuterVolumeSpecName: "kube-api-access-552z4") pod "e94ab30a-9354-4a29-80b9-6eb568d27ee9" (UID: "e94ab30a-9354-4a29-80b9-6eb568d27ee9"). InnerVolumeSpecName "kube-api-access-552z4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.364903 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e94ab30a-9354-4a29-80b9-6eb568d27ee9" (UID: "e94ab30a-9354-4a29-80b9-6eb568d27ee9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.450148 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.450406 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e94ab30a-9354-4a29-80b9-6eb568d27ee9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.450486 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-552z4\" (UniqueName: \"kubernetes.io/projected/e94ab30a-9354-4a29-80b9-6eb568d27ee9-kube-api-access-552z4\") on node \"crc\" DevicePath \"\"" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.892582 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bp825" event={"ID":"e94ab30a-9354-4a29-80b9-6eb568d27ee9","Type":"ContainerDied","Data":"3d3ced1c0d16ed550292b3c2fc8b895dde761a18c4a797d935b1be857ebaf491"} Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.892755 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bp825" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.892993 4979 scope.go:117] "RemoveContainer" containerID="e3e9d767f23e3388a560c253a764e7fc9d6370ca1ea16b62b7376591838e4d94" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.931896 4979 scope.go:117] "RemoveContainer" containerID="6255ebb20e5a32267ce7da4e3927b99bf2180929a008804f7937708d89c5b31c" Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.939574 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp825"] Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.950706 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bp825"] Dec 04 14:18:04 crc kubenswrapper[4979]: I1204 14:18:04.966452 4979 scope.go:117] "RemoveContainer" containerID="75e247e89dd3597fed99e5640d8026540fdde52a76e1ae2ae962d621b7da392f" Dec 04 14:18:05 crc kubenswrapper[4979]: E1204 14:18:05.150129 4979 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode94ab30a_9354_4a29_80b9_6eb568d27ee9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode94ab30a_9354_4a29_80b9_6eb568d27ee9.slice/crio-3d3ced1c0d16ed550292b3c2fc8b895dde761a18c4a797d935b1be857ebaf491\": RecentStats: unable to find data in memory cache]" Dec 04 14:18:06 crc kubenswrapper[4979]: I1204 14:18:06.213746 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" path="/var/lib/kubelet/pods/e94ab30a-9354-4a29-80b9-6eb568d27ee9/volumes" Dec 04 14:18:09 crc kubenswrapper[4979]: I1204 14:18:09.199692 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:18:09 crc kubenswrapper[4979]: E1204 14:18:09.200712 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:18:20 crc kubenswrapper[4979]: I1204 14:18:20.199553 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:18:20 crc kubenswrapper[4979]: E1204 14:18:20.200318 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:18:34 crc kubenswrapper[4979]: I1204 14:18:34.200277 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:18:34 crc kubenswrapper[4979]: E1204 14:18:34.200986 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:18:38 crc kubenswrapper[4979]: I1204 14:18:38.734871 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 14:18:38 crc kubenswrapper[4979]: I1204 14:18:38.735564 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mariadb-copy-data" podUID="bfae82d7-d730-474e-acb8-e0b5ef2630ab" containerName="adoption" containerID="cri-o://9c0acd65ffc63a662e9cf3545724e6de8d636eab7f51c3e02cc5072fe5758ea1" gracePeriod=30 Dec 04 14:18:48 crc kubenswrapper[4979]: I1204 14:18:48.198813 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:18:48 crc kubenswrapper[4979]: E1204 14:18:48.199638 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:19:02 crc kubenswrapper[4979]: I1204 14:19:02.199390 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:19:02 crc kubenswrapper[4979]: E1204 14:19:02.200232 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.592256 4979 generic.go:334] "Generic (PLEG): container finished" podID="bfae82d7-d730-474e-acb8-e0b5ef2630ab" containerID="9c0acd65ffc63a662e9cf3545724e6de8d636eab7f51c3e02cc5072fe5758ea1" exitCode=137 Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.592448 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"bfae82d7-d730-474e-acb8-e0b5ef2630ab","Type":"ContainerDied","Data":"9c0acd65ffc63a662e9cf3545724e6de8d636eab7f51c3e02cc5072fe5758ea1"} Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.852734 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.907842 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mariadb-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") pod \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.908002 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k48nh\" (UniqueName: \"kubernetes.io/projected/bfae82d7-d730-474e-acb8-e0b5ef2630ab-kube-api-access-k48nh\") pod \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\" (UID: \"bfae82d7-d730-474e-acb8-e0b5ef2630ab\") " Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.922730 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfae82d7-d730-474e-acb8-e0b5ef2630ab-kube-api-access-k48nh" (OuterVolumeSpecName: "kube-api-access-k48nh") pod "bfae82d7-d730-474e-acb8-e0b5ef2630ab" (UID: "bfae82d7-d730-474e-acb8-e0b5ef2630ab"). InnerVolumeSpecName "kube-api-access-k48nh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:09 crc kubenswrapper[4979]: I1204 14:19:09.972628 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327" (OuterVolumeSpecName: "mariadb-data") pod "bfae82d7-d730-474e-acb8-e0b5ef2630ab" (UID: "bfae82d7-d730-474e-acb8-e0b5ef2630ab"). InnerVolumeSpecName "pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.010969 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") on node \"crc\" " Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.011026 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k48nh\" (UniqueName: \"kubernetes.io/projected/bfae82d7-d730-474e-acb8-e0b5ef2630ab-kube-api-access-k48nh\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.120052 4979 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.120251 4979 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327") on node "crc" Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.216033 4979 reconciler_common.go:293] "Volume detached for volume \"pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-87ab06a5-8c8c-4b52-a084-8bf5ad2f2327\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.604017 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mariadb-copy-data" event={"ID":"bfae82d7-d730-474e-acb8-e0b5ef2630ab","Type":"ContainerDied","Data":"393b2e9566b3ae49806e40aa42874273f225c9769d0eb80eb9e0c243a9f252cd"} Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.604084 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mariadb-copy-data" Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.604420 4979 scope.go:117] "RemoveContainer" containerID="9c0acd65ffc63a662e9cf3545724e6de8d636eab7f51c3e02cc5072fe5758ea1" Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.635865 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 14:19:10 crc kubenswrapper[4979]: I1204 14:19:10.668199 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mariadb-copy-data"] Dec 04 14:19:11 crc kubenswrapper[4979]: I1204 14:19:11.264664 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 14:19:11 crc kubenswrapper[4979]: I1204 14:19:11.264871 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-copy-data" podUID="0856a4c9-e000-4d68-b30e-54d80b3afd79" containerName="adoption" containerID="cri-o://616e6cf4173e1da9dee11314ad17cf9d23935e33fb285e5e470c4ca84cb5e3e8" gracePeriod=30 Dec 04 14:19:12 crc kubenswrapper[4979]: I1204 14:19:12.222032 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfae82d7-d730-474e-acb8-e0b5ef2630ab" path="/var/lib/kubelet/pods/bfae82d7-d730-474e-acb8-e0b5ef2630ab/volumes" Dec 04 14:19:16 crc kubenswrapper[4979]: I1204 14:19:16.210364 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:19:16 crc kubenswrapper[4979]: E1204 14:19:16.211003 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:19:29 crc kubenswrapper[4979]: I1204 14:19:29.199412 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:19:29 crc kubenswrapper[4979]: E1204 14:19:29.200331 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:19:41 crc kubenswrapper[4979]: I1204 14:19:41.926139 4979 generic.go:334] "Generic (PLEG): container finished" podID="0856a4c9-e000-4d68-b30e-54d80b3afd79" containerID="616e6cf4173e1da9dee11314ad17cf9d23935e33fb285e5e470c4ca84cb5e3e8" exitCode=137 Dec 04 14:19:41 crc kubenswrapper[4979]: I1204 14:19:41.926234 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0856a4c9-e000-4d68-b30e-54d80b3afd79","Type":"ContainerDied","Data":"616e6cf4173e1da9dee11314ad17cf9d23935e33fb285e5e470c4ca84cb5e3e8"} Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.305050 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.406712 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45wzf\" (UniqueName: \"kubernetes.io/projected/0856a4c9-e000-4d68-b30e-54d80b3afd79-kube-api-access-45wzf\") pod \"0856a4c9-e000-4d68-b30e-54d80b3afd79\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.407605 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") pod \"0856a4c9-e000-4d68-b30e-54d80b3afd79\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.407704 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0856a4c9-e000-4d68-b30e-54d80b3afd79-ovn-data-cert\") pod \"0856a4c9-e000-4d68-b30e-54d80b3afd79\" (UID: \"0856a4c9-e000-4d68-b30e-54d80b3afd79\") " Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.412870 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0856a4c9-e000-4d68-b30e-54d80b3afd79-ovn-data-cert" (OuterVolumeSpecName: "ovn-data-cert") pod "0856a4c9-e000-4d68-b30e-54d80b3afd79" (UID: "0856a4c9-e000-4d68-b30e-54d80b3afd79"). InnerVolumeSpecName "ovn-data-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.414180 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0856a4c9-e000-4d68-b30e-54d80b3afd79-kube-api-access-45wzf" (OuterVolumeSpecName: "kube-api-access-45wzf") pod "0856a4c9-e000-4d68-b30e-54d80b3afd79" (UID: "0856a4c9-e000-4d68-b30e-54d80b3afd79"). InnerVolumeSpecName "kube-api-access-45wzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.422545 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8" (OuterVolumeSpecName: "ovn-data") pod "0856a4c9-e000-4d68-b30e-54d80b3afd79" (UID: "0856a4c9-e000-4d68-b30e-54d80b3afd79"). InnerVolumeSpecName "pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.509946 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45wzf\" (UniqueName: \"kubernetes.io/projected/0856a4c9-e000-4d68-b30e-54d80b3afd79-kube-api-access-45wzf\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.509994 4979 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") on node \"crc\" " Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.510005 4979 reconciler_common.go:293] "Volume detached for volume \"ovn-data-cert\" (UniqueName: \"kubernetes.io/secret/0856a4c9-e000-4d68-b30e-54d80b3afd79-ovn-data-cert\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.582343 4979 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.582509 4979 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8") on node "crc" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.611940 4979 reconciler_common.go:293] "Volume detached for volume \"pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-1bfec7ae-bd1b-4088-a7b1-60556d1511d8\") on node \"crc\" DevicePath \"\"" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.941920 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-copy-data" event={"ID":"0856a4c9-e000-4d68-b30e-54d80b3afd79","Type":"ContainerDied","Data":"3a8d0b0446e6b1590ec3966f962209ce05177b721c7a0956b41268cc1dfd63a2"} Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.941986 4979 scope.go:117] "RemoveContainer" containerID="616e6cf4173e1da9dee11314ad17cf9d23935e33fb285e5e470c4ca84cb5e3e8" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.942000 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-copy-data" Dec 04 14:19:42 crc kubenswrapper[4979]: I1204 14:19:42.988195 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 14:19:43 crc kubenswrapper[4979]: I1204 14:19:43.008257 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-copy-data"] Dec 04 14:19:43 crc kubenswrapper[4979]: I1204 14:19:43.200701 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:19:43 crc kubenswrapper[4979]: E1204 14:19:43.201030 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:19:44 crc kubenswrapper[4979]: I1204 14:19:44.211507 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0856a4c9-e000-4d68-b30e-54d80b3afd79" path="/var/lib/kubelet/pods/0856a4c9-e000-4d68-b30e-54d80b3afd79/volumes" Dec 04 14:19:56 crc kubenswrapper[4979]: I1204 14:19:56.210377 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:19:56 crc kubenswrapper[4979]: E1204 14:19:56.211210 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:20:11 crc kubenswrapper[4979]: I1204 14:20:11.200004 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:20:11 crc kubenswrapper[4979]: E1204 14:20:11.200888 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:20:24 crc kubenswrapper[4979]: I1204 14:20:24.199923 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:20:24 crc kubenswrapper[4979]: E1204 14:20:24.201194 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:20:35 crc kubenswrapper[4979]: I1204 14:20:35.199706 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:20:35 crc kubenswrapper[4979]: E1204 14:20:35.200588 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.317317 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lckp8/must-gather-qlk82"] Dec 04 14:20:41 crc kubenswrapper[4979]: E1204 14:20:41.318271 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0856a4c9-e000-4d68-b30e-54d80b3afd79" containerName="adoption" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318285 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="0856a4c9-e000-4d68-b30e-54d80b3afd79" containerName="adoption" Dec 04 14:20:41 crc kubenswrapper[4979]: E1204 14:20:41.318318 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="extract-utilities" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318325 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="extract-utilities" Dec 04 14:20:41 crc kubenswrapper[4979]: E1204 14:20:41.318341 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="registry-server" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318348 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="registry-server" Dec 04 14:20:41 crc kubenswrapper[4979]: E1204 14:20:41.318371 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="extract-content" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318376 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="extract-content" Dec 04 14:20:41 crc kubenswrapper[4979]: E1204 14:20:41.318390 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfae82d7-d730-474e-acb8-e0b5ef2630ab" containerName="adoption" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318395 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfae82d7-d730-474e-acb8-e0b5ef2630ab" containerName="adoption" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318603 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e94ab30a-9354-4a29-80b9-6eb568d27ee9" containerName="registry-server" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318617 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfae82d7-d730-474e-acb8-e0b5ef2630ab" containerName="adoption" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.318628 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="0856a4c9-e000-4d68-b30e-54d80b3afd79" containerName="adoption" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.319932 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.321887 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lckp8"/"kube-root-ca.crt" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.322220 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lckp8"/"default-dockercfg-l2nx5" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.322917 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lckp8"/"openshift-service-ca.crt" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.330887 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lckp8/must-gather-qlk82"] Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.450050 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2e03526b-386c-442c-91fd-d64037912796-must-gather-output\") pod \"must-gather-qlk82\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.450235 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpd47\" (UniqueName: \"kubernetes.io/projected/2e03526b-386c-442c-91fd-d64037912796-kube-api-access-rpd47\") pod \"must-gather-qlk82\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.553441 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2e03526b-386c-442c-91fd-d64037912796-must-gather-output\") pod \"must-gather-qlk82\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.553960 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2e03526b-386c-442c-91fd-d64037912796-must-gather-output\") pod \"must-gather-qlk82\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:41 crc kubenswrapper[4979]: I1204 14:20:41.554132 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpd47\" (UniqueName: \"kubernetes.io/projected/2e03526b-386c-442c-91fd-d64037912796-kube-api-access-rpd47\") pod \"must-gather-qlk82\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:42 crc kubenswrapper[4979]: I1204 14:20:42.077751 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpd47\" (UniqueName: \"kubernetes.io/projected/2e03526b-386c-442c-91fd-d64037912796-kube-api-access-rpd47\") pod \"must-gather-qlk82\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:42 crc kubenswrapper[4979]: I1204 14:20:42.237926 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:20:42 crc kubenswrapper[4979]: I1204 14:20:42.746311 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lckp8/must-gather-qlk82"] Dec 04 14:20:43 crc kubenswrapper[4979]: I1204 14:20:43.615972 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/must-gather-qlk82" event={"ID":"2e03526b-386c-442c-91fd-d64037912796","Type":"ContainerStarted","Data":"f4280bc32b89cb1c2be75dbbd92c3aa8b1ab6852ef9e40f6a7c3e6f37465d8f3"} Dec 04 14:20:49 crc kubenswrapper[4979]: I1204 14:20:49.199847 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:20:49 crc kubenswrapper[4979]: E1204 14:20:49.200810 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:20:50 crc kubenswrapper[4979]: I1204 14:20:50.687714 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/must-gather-qlk82" event={"ID":"2e03526b-386c-442c-91fd-d64037912796","Type":"ContainerStarted","Data":"5b366c969068d523774e577730fa69e4cc6a04e844cebae33ccc4dd1def7a044"} Dec 04 14:20:50 crc kubenswrapper[4979]: I1204 14:20:50.688270 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/must-gather-qlk82" event={"ID":"2e03526b-386c-442c-91fd-d64037912796","Type":"ContainerStarted","Data":"e5f057921af81b6b3c93ba28a1848b347bd3ddad12def2adc767dc9c33764be5"} Dec 04 14:20:50 crc kubenswrapper[4979]: I1204 14:20:50.713731 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lckp8/must-gather-qlk82" podStartSLOduration=2.449556853 podStartE2EDuration="9.713702423s" podCreationTimestamp="2025-12-04 14:20:41 +0000 UTC" firstStartedPulling="2025-12-04 14:20:42.75519412 +0000 UTC m=+9467.029489924" lastFinishedPulling="2025-12-04 14:20:50.01933969 +0000 UTC m=+9474.293635494" observedRunningTime="2025-12-04 14:20:50.706018738 +0000 UTC m=+9474.980314552" watchObservedRunningTime="2025-12-04 14:20:50.713702423 +0000 UTC m=+9474.987998227" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.047241 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lckp8/crc-debug-6gxv4"] Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.049380 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.086403 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48mkf\" (UniqueName: \"kubernetes.io/projected/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-kube-api-access-48mkf\") pod \"crc-debug-6gxv4\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.086483 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-host\") pod \"crc-debug-6gxv4\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.189201 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48mkf\" (UniqueName: \"kubernetes.io/projected/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-kube-api-access-48mkf\") pod \"crc-debug-6gxv4\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.189278 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-host\") pod \"crc-debug-6gxv4\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.189510 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-host\") pod \"crc-debug-6gxv4\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.214013 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48mkf\" (UniqueName: \"kubernetes.io/projected/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-kube-api-access-48mkf\") pod \"crc-debug-6gxv4\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.372546 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:20:58 crc kubenswrapper[4979]: I1204 14:20:58.783659 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" event={"ID":"8b5466d6-e89b-4634-8c81-19bfe71ae7ce","Type":"ContainerStarted","Data":"5af099a2533b0eadc1ee63742dca67e55175b4b2e303e07008ae5afc67d82197"} Dec 04 14:21:04 crc kubenswrapper[4979]: I1204 14:21:04.199588 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:21:04 crc kubenswrapper[4979]: I1204 14:21:04.846668 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"673e4ca9deca044cfcadec94d360ec56cdf6d25ae2da6226a912caa24ffed270"} Dec 04 14:21:24 crc kubenswrapper[4979]: E1204 14:21:24.174242 4979 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 04 14:21:24 crc kubenswrapper[4979]: E1204 14:21:24.174981 4979 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-48mkf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-6gxv4_openshift-must-gather-lckp8(8b5466d6-e89b-4634-8c81-19bfe71ae7ce): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 04 14:21:24 crc kubenswrapper[4979]: E1204 14:21:24.176478 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" podUID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" Dec 04 14:21:25 crc kubenswrapper[4979]: E1204 14:21:25.056211 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" podUID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" Dec 04 14:21:39 crc kubenswrapper[4979]: I1204 14:21:39.198983 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" event={"ID":"8b5466d6-e89b-4634-8c81-19bfe71ae7ce","Type":"ContainerStarted","Data":"26d9c462b24463f90be003be2b087a55d898e3ee041c6f4f86e7be19d7cd2efa"} Dec 04 14:22:06 crc kubenswrapper[4979]: I1204 14:22:06.507766 4979 generic.go:334] "Generic (PLEG): container finished" podID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" containerID="26d9c462b24463f90be003be2b087a55d898e3ee041c6f4f86e7be19d7cd2efa" exitCode=0 Dec 04 14:22:06 crc kubenswrapper[4979]: I1204 14:22:06.507822 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" event={"ID":"8b5466d6-e89b-4634-8c81-19bfe71ae7ce","Type":"ContainerDied","Data":"26d9c462b24463f90be003be2b087a55d898e3ee041c6f4f86e7be19d7cd2efa"} Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.647837 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.681033 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lckp8/crc-debug-6gxv4"] Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.692188 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lckp8/crc-debug-6gxv4"] Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.804818 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-host\") pod \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.804985 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-host" (OuterVolumeSpecName: "host") pod "8b5466d6-e89b-4634-8c81-19bfe71ae7ce" (UID: "8b5466d6-e89b-4634-8c81-19bfe71ae7ce"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.805026 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48mkf\" (UniqueName: \"kubernetes.io/projected/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-kube-api-access-48mkf\") pod \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\" (UID: \"8b5466d6-e89b-4634-8c81-19bfe71ae7ce\") " Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.805856 4979 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-host\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.816576 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-kube-api-access-48mkf" (OuterVolumeSpecName: "kube-api-access-48mkf") pod "8b5466d6-e89b-4634-8c81-19bfe71ae7ce" (UID: "8b5466d6-e89b-4634-8c81-19bfe71ae7ce"). InnerVolumeSpecName "kube-api-access-48mkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:07 crc kubenswrapper[4979]: I1204 14:22:07.908537 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48mkf\" (UniqueName: \"kubernetes.io/projected/8b5466d6-e89b-4634-8c81-19bfe71ae7ce-kube-api-access-48mkf\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.212130 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" path="/var/lib/kubelet/pods/8b5466d6-e89b-4634-8c81-19bfe71ae7ce/volumes" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.529279 4979 scope.go:117] "RemoveContainer" containerID="26d9c462b24463f90be003be2b087a55d898e3ee041c6f4f86e7be19d7cd2efa" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.529341 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-6gxv4" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.862837 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lckp8/crc-debug-gzxlh"] Dec 04 14:22:08 crc kubenswrapper[4979]: E1204 14:22:08.863635 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" containerName="container-00" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.863649 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" containerName="container-00" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.863845 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b5466d6-e89b-4634-8c81-19bfe71ae7ce" containerName="container-00" Dec 04 14:22:08 crc kubenswrapper[4979]: I1204 14:22:08.864622 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.031783 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e46a3fb-b710-4e60-b11e-3e396743ac8b-host\") pod \"crc-debug-gzxlh\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.032179 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8vnv\" (UniqueName: \"kubernetes.io/projected/7e46a3fb-b710-4e60-b11e-3e396743ac8b-kube-api-access-p8vnv\") pod \"crc-debug-gzxlh\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.134508 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8vnv\" (UniqueName: \"kubernetes.io/projected/7e46a3fb-b710-4e60-b11e-3e396743ac8b-kube-api-access-p8vnv\") pod \"crc-debug-gzxlh\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.134727 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e46a3fb-b710-4e60-b11e-3e396743ac8b-host\") pod \"crc-debug-gzxlh\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.134958 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e46a3fb-b710-4e60-b11e-3e396743ac8b-host\") pod \"crc-debug-gzxlh\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.153775 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8vnv\" (UniqueName: \"kubernetes.io/projected/7e46a3fb-b710-4e60-b11e-3e396743ac8b-kube-api-access-p8vnv\") pod \"crc-debug-gzxlh\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.185337 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:09 crc kubenswrapper[4979]: W1204 14:22:09.215423 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e46a3fb_b710_4e60_b11e_3e396743ac8b.slice/crio-7430581d596fe497a987d38bc4f56fd1dbca9cde34334d9972d8f0d244ceebc7 WatchSource:0}: Error finding container 7430581d596fe497a987d38bc4f56fd1dbca9cde34334d9972d8f0d244ceebc7: Status 404 returned error can't find the container with id 7430581d596fe497a987d38bc4f56fd1dbca9cde34334d9972d8f0d244ceebc7 Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.543965 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" event={"ID":"7e46a3fb-b710-4e60-b11e-3e396743ac8b","Type":"ContainerStarted","Data":"0eddb6999cd2eecd2c19230ba645dff92757211e4e752961c99c2f14f74b9c13"} Dec 04 14:22:09 crc kubenswrapper[4979]: I1204 14:22:09.544442 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" event={"ID":"7e46a3fb-b710-4e60-b11e-3e396743ac8b","Type":"ContainerStarted","Data":"7430581d596fe497a987d38bc4f56fd1dbca9cde34334d9972d8f0d244ceebc7"} Dec 04 14:22:10 crc kubenswrapper[4979]: I1204 14:22:10.557741 4979 generic.go:334] "Generic (PLEG): container finished" podID="7e46a3fb-b710-4e60-b11e-3e396743ac8b" containerID="0eddb6999cd2eecd2c19230ba645dff92757211e4e752961c99c2f14f74b9c13" exitCode=1 Dec 04 14:22:10 crc kubenswrapper[4979]: I1204 14:22:10.557808 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" event={"ID":"7e46a3fb-b710-4e60-b11e-3e396743ac8b","Type":"ContainerDied","Data":"0eddb6999cd2eecd2c19230ba645dff92757211e4e752961c99c2f14f74b9c13"} Dec 04 14:22:10 crc kubenswrapper[4979]: I1204 14:22:10.610655 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lckp8/crc-debug-gzxlh"] Dec 04 14:22:10 crc kubenswrapper[4979]: I1204 14:22:10.621688 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lckp8/crc-debug-gzxlh"] Dec 04 14:22:11 crc kubenswrapper[4979]: I1204 14:22:11.794385 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:22:11 crc kubenswrapper[4979]: I1204 14:22:11.917694 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8vnv\" (UniqueName: \"kubernetes.io/projected/7e46a3fb-b710-4e60-b11e-3e396743ac8b-kube-api-access-p8vnv\") pod \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " Dec 04 14:22:11 crc kubenswrapper[4979]: I1204 14:22:11.918032 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e46a3fb-b710-4e60-b11e-3e396743ac8b-host\") pod \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\" (UID: \"7e46a3fb-b710-4e60-b11e-3e396743ac8b\") " Dec 04 14:22:11 crc kubenswrapper[4979]: I1204 14:22:11.918508 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7e46a3fb-b710-4e60-b11e-3e396743ac8b-host" (OuterVolumeSpecName: "host") pod "7e46a3fb-b710-4e60-b11e-3e396743ac8b" (UID: "7e46a3fb-b710-4e60-b11e-3e396743ac8b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 04 14:22:11 crc kubenswrapper[4979]: I1204 14:22:11.918863 4979 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7e46a3fb-b710-4e60-b11e-3e396743ac8b-host\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:11 crc kubenswrapper[4979]: I1204 14:22:11.926669 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e46a3fb-b710-4e60-b11e-3e396743ac8b-kube-api-access-p8vnv" (OuterVolumeSpecName: "kube-api-access-p8vnv") pod "7e46a3fb-b710-4e60-b11e-3e396743ac8b" (UID: "7e46a3fb-b710-4e60-b11e-3e396743ac8b"). InnerVolumeSpecName "kube-api-access-p8vnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:22:12 crc kubenswrapper[4979]: I1204 14:22:12.023463 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8vnv\" (UniqueName: \"kubernetes.io/projected/7e46a3fb-b710-4e60-b11e-3e396743ac8b-kube-api-access-p8vnv\") on node \"crc\" DevicePath \"\"" Dec 04 14:22:12 crc kubenswrapper[4979]: I1204 14:22:12.249192 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e46a3fb-b710-4e60-b11e-3e396743ac8b" path="/var/lib/kubelet/pods/7e46a3fb-b710-4e60-b11e-3e396743ac8b/volumes" Dec 04 14:22:12 crc kubenswrapper[4979]: I1204 14:22:12.581978 4979 scope.go:117] "RemoveContainer" containerID="0eddb6999cd2eecd2c19230ba645dff92757211e4e752961c99c2f14f74b9c13" Dec 04 14:22:12 crc kubenswrapper[4979]: I1204 14:22:12.582188 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/crc-debug-gzxlh" Dec 04 14:23:28 crc kubenswrapper[4979]: I1204 14:23:28.040470 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:23:28 crc kubenswrapper[4979]: I1204 14:23:28.041109 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:23:58 crc kubenswrapper[4979]: I1204 14:23:58.041209 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:23:58 crc kubenswrapper[4979]: I1204 14:23:58.041858 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.041490 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.042101 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.042148 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.043068 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"673e4ca9deca044cfcadec94d360ec56cdf6d25ae2da6226a912caa24ffed270"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.043128 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://673e4ca9deca044cfcadec94d360ec56cdf6d25ae2da6226a912caa24ffed270" gracePeriod=600 Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.256682 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hqnw9"] Dec 04 14:24:28 crc kubenswrapper[4979]: E1204 14:24:28.258130 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e46a3fb-b710-4e60-b11e-3e396743ac8b" containerName="container-00" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.258223 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e46a3fb-b710-4e60-b11e-3e396743ac8b" containerName="container-00" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.258525 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e46a3fb-b710-4e60-b11e-3e396743ac8b" containerName="container-00" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.260074 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.272028 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqnw9"] Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.439730 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx5vp\" (UniqueName: \"kubernetes.io/projected/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-kube-api-access-xx5vp\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.440198 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-utilities\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.440417 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-catalog-content\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.542463 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx5vp\" (UniqueName: \"kubernetes.io/projected/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-kube-api-access-xx5vp\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.542943 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-utilities\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.543407 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-utilities\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.543904 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-catalog-content\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.544772 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-catalog-content\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.564694 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx5vp\" (UniqueName: \"kubernetes.io/projected/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-kube-api-access-xx5vp\") pod \"redhat-operators-hqnw9\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:28 crc kubenswrapper[4979]: I1204 14:24:28.595264 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:29 crc kubenswrapper[4979]: I1204 14:24:29.086476 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="673e4ca9deca044cfcadec94d360ec56cdf6d25ae2da6226a912caa24ffed270" exitCode=0 Dec 04 14:24:29 crc kubenswrapper[4979]: I1204 14:24:29.086555 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"673e4ca9deca044cfcadec94d360ec56cdf6d25ae2da6226a912caa24ffed270"} Dec 04 14:24:29 crc kubenswrapper[4979]: I1204 14:24:29.086991 4979 scope.go:117] "RemoveContainer" containerID="19c55908e5f237bc11d9425ddd6f82380aa74baa05309da2b3312c8f829537a1" Dec 04 14:24:29 crc kubenswrapper[4979]: I1204 14:24:29.135139 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hqnw9"] Dec 04 14:24:30 crc kubenswrapper[4979]: I1204 14:24:30.099810 4979 generic.go:334] "Generic (PLEG): container finished" podID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerID="103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87" exitCode=0 Dec 04 14:24:30 crc kubenswrapper[4979]: I1204 14:24:30.099923 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerDied","Data":"103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87"} Dec 04 14:24:30 crc kubenswrapper[4979]: I1204 14:24:30.101361 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerStarted","Data":"acad50529d1b2de4c7a50ebcba2923f3ea180637a2da1059c8a1cf7f62ce2b24"} Dec 04 14:24:30 crc kubenswrapper[4979]: I1204 14:24:30.102229 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:24:30 crc kubenswrapper[4979]: I1204 14:24:30.103788 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0"} Dec 04 14:24:33 crc kubenswrapper[4979]: I1204 14:24:33.139281 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerStarted","Data":"14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55"} Dec 04 14:24:37 crc kubenswrapper[4979]: I1204 14:24:37.178716 4979 generic.go:334] "Generic (PLEG): container finished" podID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerID="14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55" exitCode=0 Dec 04 14:24:37 crc kubenswrapper[4979]: I1204 14:24:37.178772 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerDied","Data":"14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55"} Dec 04 14:24:38 crc kubenswrapper[4979]: I1204 14:24:38.190324 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerStarted","Data":"eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e"} Dec 04 14:24:38 crc kubenswrapper[4979]: I1204 14:24:38.216075 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hqnw9" podStartSLOduration=2.730112306 podStartE2EDuration="10.216052882s" podCreationTimestamp="2025-12-04 14:24:28 +0000 UTC" firstStartedPulling="2025-12-04 14:24:30.102009229 +0000 UTC m=+9694.376305033" lastFinishedPulling="2025-12-04 14:24:37.587949805 +0000 UTC m=+9701.862245609" observedRunningTime="2025-12-04 14:24:38.20740711 +0000 UTC m=+9702.481702924" watchObservedRunningTime="2025-12-04 14:24:38.216052882 +0000 UTC m=+9702.490348696" Dec 04 14:24:38 crc kubenswrapper[4979]: I1204 14:24:38.595648 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:38 crc kubenswrapper[4979]: I1204 14:24:38.595849 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:39 crc kubenswrapper[4979]: I1204 14:24:39.643706 4979 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hqnw9" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="registry-server" probeResult="failure" output=< Dec 04 14:24:39 crc kubenswrapper[4979]: timeout: failed to connect service ":50051" within 1s Dec 04 14:24:39 crc kubenswrapper[4979]: > Dec 04 14:24:48 crc kubenswrapper[4979]: I1204 14:24:48.835932 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:48 crc kubenswrapper[4979]: I1204 14:24:48.881877 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:49 crc kubenswrapper[4979]: I1204 14:24:49.076030 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqnw9"] Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.333554 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hqnw9" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="registry-server" containerID="cri-o://eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e" gracePeriod=2 Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.856056 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.963960 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-catalog-content\") pod \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.964726 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xx5vp\" (UniqueName: \"kubernetes.io/projected/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-kube-api-access-xx5vp\") pod \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.965039 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-utilities\") pod \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\" (UID: \"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153\") " Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.966753 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-utilities" (OuterVolumeSpecName: "utilities") pod "e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" (UID: "e59ea4c9-5c98-4d2a-92c0-8bafb33f3153"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:24:50 crc kubenswrapper[4979]: I1204 14:24:50.973028 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-kube-api-access-xx5vp" (OuterVolumeSpecName: "kube-api-access-xx5vp") pod "e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" (UID: "e59ea4c9-5c98-4d2a-92c0-8bafb33f3153"). InnerVolumeSpecName "kube-api-access-xx5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.067705 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xx5vp\" (UniqueName: \"kubernetes.io/projected/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-kube-api-access-xx5vp\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.067740 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.086178 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" (UID: "e59ea4c9-5c98-4d2a-92c0-8bafb33f3153"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.170387 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.350085 4979 generic.go:334] "Generic (PLEG): container finished" podID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerID="eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e" exitCode=0 Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.350169 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hqnw9" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.350213 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerDied","Data":"eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e"} Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.350701 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hqnw9" event={"ID":"e59ea4c9-5c98-4d2a-92c0-8bafb33f3153","Type":"ContainerDied","Data":"acad50529d1b2de4c7a50ebcba2923f3ea180637a2da1059c8a1cf7f62ce2b24"} Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.350754 4979 scope.go:117] "RemoveContainer" containerID="eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.388034 4979 scope.go:117] "RemoveContainer" containerID="14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.390094 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hqnw9"] Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.401891 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hqnw9"] Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.407789 4979 scope.go:117] "RemoveContainer" containerID="103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.460607 4979 scope.go:117] "RemoveContainer" containerID="eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e" Dec 04 14:24:51 crc kubenswrapper[4979]: E1204 14:24:51.463217 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e\": container with ID starting with eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e not found: ID does not exist" containerID="eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.463287 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e"} err="failed to get container status \"eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e\": rpc error: code = NotFound desc = could not find container \"eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e\": container with ID starting with eeab393417e55542b5e0b4d5a17c0920a6a49bf69ea865c8edc887786644751e not found: ID does not exist" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.463333 4979 scope.go:117] "RemoveContainer" containerID="14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55" Dec 04 14:24:51 crc kubenswrapper[4979]: E1204 14:24:51.463712 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55\": container with ID starting with 14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55 not found: ID does not exist" containerID="14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.463749 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55"} err="failed to get container status \"14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55\": rpc error: code = NotFound desc = could not find container \"14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55\": container with ID starting with 14d1b2ff5df46f57d940db00306c883467e411850eedea0f9af44f7918327d55 not found: ID does not exist" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.463778 4979 scope.go:117] "RemoveContainer" containerID="103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87" Dec 04 14:24:51 crc kubenswrapper[4979]: E1204 14:24:51.464135 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87\": container with ID starting with 103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87 not found: ID does not exist" containerID="103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87" Dec 04 14:24:51 crc kubenswrapper[4979]: I1204 14:24:51.464166 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87"} err="failed to get container status \"103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87\": rpc error: code = NotFound desc = could not find container \"103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87\": container with ID starting with 103159cef2a39128e921a35b84ce934a07fa876ea1ea2abbd7cd58c4537cda87 not found: ID does not exist" Dec 04 14:24:52 crc kubenswrapper[4979]: I1204 14:24:52.216800 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" path="/var/lib/kubelet/pods/e59ea4c9-5c98-4d2a-92c0-8bafb33f3153/volumes" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.525335 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-k5s4n"] Dec 04 14:25:04 crc kubenswrapper[4979]: E1204 14:25:04.526275 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="extract-content" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.526289 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="extract-content" Dec 04 14:25:04 crc kubenswrapper[4979]: E1204 14:25:04.526331 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="registry-server" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.526338 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="registry-server" Dec 04 14:25:04 crc kubenswrapper[4979]: E1204 14:25:04.526357 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="extract-utilities" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.526363 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="extract-utilities" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.526587 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59ea4c9-5c98-4d2a-92c0-8bafb33f3153" containerName="registry-server" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.528343 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.555229 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k5s4n"] Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.680185 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-utilities\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.680271 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx9r5\" (UniqueName: \"kubernetes.io/projected/006b9d3b-f389-46ae-a38e-ed08e9b42fce-kube-api-access-hx9r5\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.680732 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-catalog-content\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.782873 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-catalog-content\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.783263 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-utilities\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.783358 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx9r5\" (UniqueName: \"kubernetes.io/projected/006b9d3b-f389-46ae-a38e-ed08e9b42fce-kube-api-access-hx9r5\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.783371 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-catalog-content\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.783645 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-utilities\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.808420 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx9r5\" (UniqueName: \"kubernetes.io/projected/006b9d3b-f389-46ae-a38e-ed08e9b42fce-kube-api-access-hx9r5\") pod \"certified-operators-k5s4n\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:04 crc kubenswrapper[4979]: I1204 14:25:04.860336 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:05 crc kubenswrapper[4979]: I1204 14:25:05.461133 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-k5s4n"] Dec 04 14:25:05 crc kubenswrapper[4979]: I1204 14:25:05.518331 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5s4n" event={"ID":"006b9d3b-f389-46ae-a38e-ed08e9b42fce","Type":"ContainerStarted","Data":"86a2af1c2c250f6cf04dad1e8b77a13d747bf3d69db1f6ed1f793b4ebf422bc5"} Dec 04 14:25:06 crc kubenswrapper[4979]: I1204 14:25:06.530649 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5s4n" event={"ID":"006b9d3b-f389-46ae-a38e-ed08e9b42fce","Type":"ContainerDied","Data":"224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7"} Dec 04 14:25:06 crc kubenswrapper[4979]: I1204 14:25:06.530485 4979 generic.go:334] "Generic (PLEG): container finished" podID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerID="224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7" exitCode=0 Dec 04 14:25:08 crc kubenswrapper[4979]: I1204 14:25:08.549591 4979 generic.go:334] "Generic (PLEG): container finished" podID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerID="0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc" exitCode=0 Dec 04 14:25:08 crc kubenswrapper[4979]: I1204 14:25:08.549734 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5s4n" event={"ID":"006b9d3b-f389-46ae-a38e-ed08e9b42fce","Type":"ContainerDied","Data":"0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc"} Dec 04 14:25:09 crc kubenswrapper[4979]: I1204 14:25:09.562006 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5s4n" event={"ID":"006b9d3b-f389-46ae-a38e-ed08e9b42fce","Type":"ContainerStarted","Data":"b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c"} Dec 04 14:25:09 crc kubenswrapper[4979]: I1204 14:25:09.578863 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-k5s4n" podStartSLOduration=3.059642545 podStartE2EDuration="5.578843253s" podCreationTimestamp="2025-12-04 14:25:04 +0000 UTC" firstStartedPulling="2025-12-04 14:25:06.533723207 +0000 UTC m=+9730.808019011" lastFinishedPulling="2025-12-04 14:25:09.052923915 +0000 UTC m=+9733.327219719" observedRunningTime="2025-12-04 14:25:09.577546078 +0000 UTC m=+9733.851841892" watchObservedRunningTime="2025-12-04 14:25:09.578843253 +0000 UTC m=+9733.853139067" Dec 04 14:25:10 crc kubenswrapper[4979]: I1204 14:25:10.911601 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zhm6c"] Dec 04 14:25:10 crc kubenswrapper[4979]: I1204 14:25:10.914419 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:10 crc kubenswrapper[4979]: I1204 14:25:10.925935 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zhm6c"] Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.022792 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm84t\" (UniqueName: \"kubernetes.io/projected/74114e27-f36b-47b2-8088-cae23a80ddae-kube-api-access-fm84t\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.023219 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-utilities\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.023434 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-catalog-content\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.124970 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-catalog-content\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.125048 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm84t\" (UniqueName: \"kubernetes.io/projected/74114e27-f36b-47b2-8088-cae23a80ddae-kube-api-access-fm84t\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.125174 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-utilities\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.125757 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-catalog-content\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.128929 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-utilities\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.166192 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm84t\" (UniqueName: \"kubernetes.io/projected/74114e27-f36b-47b2-8088-cae23a80ddae-kube-api-access-fm84t\") pod \"community-operators-zhm6c\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.235801 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:11 crc kubenswrapper[4979]: W1204 14:25:11.764241 4979 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74114e27_f36b_47b2_8088_cae23a80ddae.slice/crio-f8068b5c58f195a383b9eca36bb0e84ab76807ee46a95c34e1e87a573d0df44c WatchSource:0}: Error finding container f8068b5c58f195a383b9eca36bb0e84ab76807ee46a95c34e1e87a573d0df44c: Status 404 returned error can't find the container with id f8068b5c58f195a383b9eca36bb0e84ab76807ee46a95c34e1e87a573d0df44c Dec 04 14:25:11 crc kubenswrapper[4979]: I1204 14:25:11.764921 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zhm6c"] Dec 04 14:25:12 crc kubenswrapper[4979]: I1204 14:25:12.591678 4979 generic.go:334] "Generic (PLEG): container finished" podID="74114e27-f36b-47b2-8088-cae23a80ddae" containerID="827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151" exitCode=0 Dec 04 14:25:12 crc kubenswrapper[4979]: I1204 14:25:12.591778 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerDied","Data":"827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151"} Dec 04 14:25:12 crc kubenswrapper[4979]: I1204 14:25:12.592008 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerStarted","Data":"f8068b5c58f195a383b9eca36bb0e84ab76807ee46a95c34e1e87a573d0df44c"} Dec 04 14:25:14 crc kubenswrapper[4979]: I1204 14:25:14.616020 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerStarted","Data":"367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1"} Dec 04 14:25:14 crc kubenswrapper[4979]: I1204 14:25:14.861218 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:14 crc kubenswrapper[4979]: I1204 14:25:14.861269 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:14 crc kubenswrapper[4979]: I1204 14:25:14.912983 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:15 crc kubenswrapper[4979]: I1204 14:25:15.671240 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:16 crc kubenswrapper[4979]: I1204 14:25:16.096279 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k5s4n"] Dec 04 14:25:16 crc kubenswrapper[4979]: I1204 14:25:16.635717 4979 generic.go:334] "Generic (PLEG): container finished" podID="74114e27-f36b-47b2-8088-cae23a80ddae" containerID="367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1" exitCode=0 Dec 04 14:25:16 crc kubenswrapper[4979]: I1204 14:25:16.636421 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerDied","Data":"367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1"} Dec 04 14:25:17 crc kubenswrapper[4979]: I1204 14:25:17.647226 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerStarted","Data":"a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58"} Dec 04 14:25:17 crc kubenswrapper[4979]: I1204 14:25:17.647369 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-k5s4n" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="registry-server" containerID="cri-o://b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c" gracePeriod=2 Dec 04 14:25:17 crc kubenswrapper[4979]: I1204 14:25:17.675846 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zhm6c" podStartSLOduration=3.23504976 podStartE2EDuration="7.675825039s" podCreationTimestamp="2025-12-04 14:25:10 +0000 UTC" firstStartedPulling="2025-12-04 14:25:12.596337039 +0000 UTC m=+9736.870632853" lastFinishedPulling="2025-12-04 14:25:17.037112328 +0000 UTC m=+9741.311408132" observedRunningTime="2025-12-04 14:25:17.671225166 +0000 UTC m=+9741.945520970" watchObservedRunningTime="2025-12-04 14:25:17.675825039 +0000 UTC m=+9741.950120843" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.180728 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.285099 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx9r5\" (UniqueName: \"kubernetes.io/projected/006b9d3b-f389-46ae-a38e-ed08e9b42fce-kube-api-access-hx9r5\") pod \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.285195 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-catalog-content\") pod \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.285534 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-utilities\") pod \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\" (UID: \"006b9d3b-f389-46ae-a38e-ed08e9b42fce\") " Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.287246 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-utilities" (OuterVolumeSpecName: "utilities") pod "006b9d3b-f389-46ae-a38e-ed08e9b42fce" (UID: "006b9d3b-f389-46ae-a38e-ed08e9b42fce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.292028 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006b9d3b-f389-46ae-a38e-ed08e9b42fce-kube-api-access-hx9r5" (OuterVolumeSpecName: "kube-api-access-hx9r5") pod "006b9d3b-f389-46ae-a38e-ed08e9b42fce" (UID: "006b9d3b-f389-46ae-a38e-ed08e9b42fce"). InnerVolumeSpecName "kube-api-access-hx9r5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.332982 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "006b9d3b-f389-46ae-a38e-ed08e9b42fce" (UID: "006b9d3b-f389-46ae-a38e-ed08e9b42fce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.388704 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.388744 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx9r5\" (UniqueName: \"kubernetes.io/projected/006b9d3b-f389-46ae-a38e-ed08e9b42fce-kube-api-access-hx9r5\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.388757 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006b9d3b-f389-46ae-a38e-ed08e9b42fce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.660253 4979 generic.go:334] "Generic (PLEG): container finished" podID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerID="b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c" exitCode=0 Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.660333 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-k5s4n" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.660332 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5s4n" event={"ID":"006b9d3b-f389-46ae-a38e-ed08e9b42fce","Type":"ContainerDied","Data":"b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c"} Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.660911 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-k5s4n" event={"ID":"006b9d3b-f389-46ae-a38e-ed08e9b42fce","Type":"ContainerDied","Data":"86a2af1c2c250f6cf04dad1e8b77a13d747bf3d69db1f6ed1f793b4ebf422bc5"} Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.660955 4979 scope.go:117] "RemoveContainer" containerID="b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.692530 4979 scope.go:117] "RemoveContainer" containerID="0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.703091 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-k5s4n"] Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.713837 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-k5s4n"] Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.729103 4979 scope.go:117] "RemoveContainer" containerID="224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.768620 4979 scope.go:117] "RemoveContainer" containerID="b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c" Dec 04 14:25:18 crc kubenswrapper[4979]: E1204 14:25:18.769166 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c\": container with ID starting with b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c not found: ID does not exist" containerID="b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.769217 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c"} err="failed to get container status \"b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c\": rpc error: code = NotFound desc = could not find container \"b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c\": container with ID starting with b0f9ab9e57bc13ed3e70a6b032fb61cbf00a12085afac546fe3d65cab3ec0a2c not found: ID does not exist" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.769249 4979 scope.go:117] "RemoveContainer" containerID="0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc" Dec 04 14:25:18 crc kubenswrapper[4979]: E1204 14:25:18.770090 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc\": container with ID starting with 0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc not found: ID does not exist" containerID="0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.770132 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc"} err="failed to get container status \"0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc\": rpc error: code = NotFound desc = could not find container \"0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc\": container with ID starting with 0ccacd0ede06825e8d09aff47cf63b113c895016b0b1b4c99efb1c79eeaa79cc not found: ID does not exist" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.770159 4979 scope.go:117] "RemoveContainer" containerID="224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7" Dec 04 14:25:18 crc kubenswrapper[4979]: E1204 14:25:18.770569 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7\": container with ID starting with 224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7 not found: ID does not exist" containerID="224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7" Dec 04 14:25:18 crc kubenswrapper[4979]: I1204 14:25:18.770617 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7"} err="failed to get container status \"224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7\": rpc error: code = NotFound desc = could not find container \"224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7\": container with ID starting with 224406574ab93fd8cb3b97a2aa148c732b63f196038ae211fed986ee30ea88b7 not found: ID does not exist" Dec 04 14:25:20 crc kubenswrapper[4979]: I1204 14:25:20.213668 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" path="/var/lib/kubelet/pods/006b9d3b-f389-46ae-a38e-ed08e9b42fce/volumes" Dec 04 14:25:21 crc kubenswrapper[4979]: I1204 14:25:21.236966 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:21 crc kubenswrapper[4979]: I1204 14:25:21.237434 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:21 crc kubenswrapper[4979]: I1204 14:25:21.306524 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:31 crc kubenswrapper[4979]: I1204 14:25:31.290464 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:31 crc kubenswrapper[4979]: I1204 14:25:31.366523 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zhm6c"] Dec 04 14:25:31 crc kubenswrapper[4979]: I1204 14:25:31.802698 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zhm6c" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="registry-server" containerID="cri-o://a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58" gracePeriod=2 Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.372614 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.538455 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fm84t\" (UniqueName: \"kubernetes.io/projected/74114e27-f36b-47b2-8088-cae23a80ddae-kube-api-access-fm84t\") pod \"74114e27-f36b-47b2-8088-cae23a80ddae\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.538704 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-utilities\") pod \"74114e27-f36b-47b2-8088-cae23a80ddae\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.538780 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-catalog-content\") pod \"74114e27-f36b-47b2-8088-cae23a80ddae\" (UID: \"74114e27-f36b-47b2-8088-cae23a80ddae\") " Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.539841 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-utilities" (OuterVolumeSpecName: "utilities") pod "74114e27-f36b-47b2-8088-cae23a80ddae" (UID: "74114e27-f36b-47b2-8088-cae23a80ddae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.545671 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74114e27-f36b-47b2-8088-cae23a80ddae-kube-api-access-fm84t" (OuterVolumeSpecName: "kube-api-access-fm84t") pod "74114e27-f36b-47b2-8088-cae23a80ddae" (UID: "74114e27-f36b-47b2-8088-cae23a80ddae"). InnerVolumeSpecName "kube-api-access-fm84t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.605804 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74114e27-f36b-47b2-8088-cae23a80ddae" (UID: "74114e27-f36b-47b2-8088-cae23a80ddae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.641561 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fm84t\" (UniqueName: \"kubernetes.io/projected/74114e27-f36b-47b2-8088-cae23a80ddae-kube-api-access-fm84t\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.641615 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.641628 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74114e27-f36b-47b2-8088-cae23a80ddae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.815796 4979 generic.go:334] "Generic (PLEG): container finished" podID="74114e27-f36b-47b2-8088-cae23a80ddae" containerID="a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58" exitCode=0 Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.815873 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerDied","Data":"a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58"} Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.815951 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zhm6c" event={"ID":"74114e27-f36b-47b2-8088-cae23a80ddae","Type":"ContainerDied","Data":"f8068b5c58f195a383b9eca36bb0e84ab76807ee46a95c34e1e87a573d0df44c"} Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.815970 4979 scope.go:117] "RemoveContainer" containerID="a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.815889 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zhm6c" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.851189 4979 scope.go:117] "RemoveContainer" containerID="367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.855766 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zhm6c"] Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.866461 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zhm6c"] Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.884500 4979 scope.go:117] "RemoveContainer" containerID="827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.905249 4979 scope.go:117] "RemoveContainer" containerID="a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58" Dec 04 14:25:32 crc kubenswrapper[4979]: E1204 14:25:32.905717 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58\": container with ID starting with a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58 not found: ID does not exist" containerID="a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.905772 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58"} err="failed to get container status \"a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58\": rpc error: code = NotFound desc = could not find container \"a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58\": container with ID starting with a8974ef6637e5ef18fb2157cd38f50dfbba8494713fb710085c88a3815cdcd58 not found: ID does not exist" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.905819 4979 scope.go:117] "RemoveContainer" containerID="367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1" Dec 04 14:25:32 crc kubenswrapper[4979]: E1204 14:25:32.906141 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1\": container with ID starting with 367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1 not found: ID does not exist" containerID="367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.906170 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1"} err="failed to get container status \"367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1\": rpc error: code = NotFound desc = could not find container \"367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1\": container with ID starting with 367d5604ca44fee178831454f904e2f57029fd586461911b9a3a306624c3baa1 not found: ID does not exist" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.906192 4979 scope.go:117] "RemoveContainer" containerID="827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151" Dec 04 14:25:32 crc kubenswrapper[4979]: E1204 14:25:32.906732 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151\": container with ID starting with 827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151 not found: ID does not exist" containerID="827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151" Dec 04 14:25:32 crc kubenswrapper[4979]: I1204 14:25:32.906766 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151"} err="failed to get container status \"827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151\": rpc error: code = NotFound desc = could not find container \"827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151\": container with ID starting with 827a6c80a6b1121eb3cbc6f830ecf59abe5d3dffe65836c196b013405cee6151 not found: ID does not exist" Dec 04 14:25:34 crc kubenswrapper[4979]: I1204 14:25:34.214659 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" path="/var/lib/kubelet/pods/74114e27-f36b-47b2-8088-cae23a80ddae/volumes" Dec 04 14:26:58 crc kubenswrapper[4979]: I1204 14:26:58.041335 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:26:58 crc kubenswrapper[4979]: I1204 14:26:58.041883 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:27:28 crc kubenswrapper[4979]: I1204 14:27:28.040658 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:27:28 crc kubenswrapper[4979]: I1204 14:27:28.042776 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:27:30 crc kubenswrapper[4979]: I1204 14:27:30.362526 4979 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-n6rcx container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.1.131:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:27:30 crc kubenswrapper[4979]: I1204 14:27:30.364848 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" podUID="eeec64e3-82f9-4370-8621-fa7963371a38" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.1.131:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:27:30 crc kubenswrapper[4979]: I1204 14:27:30.362550 4979 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-n6rcx container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.1.131:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:27:30 crc kubenswrapper[4979]: I1204 14:27:30.365427 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-n6rcx" podUID="eeec64e3-82f9-4370-8621-fa7963371a38" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.1.131:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:27:41 crc kubenswrapper[4979]: I1204 14:27:41.774258 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="462e9872-5a8f-4fd0-bd81-cf2ec4471b60" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 04 14:27:42 crc kubenswrapper[4979]: I1204 14:27:42.552460 4979 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-fnchd container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:27:42 crc kubenswrapper[4979]: I1204 14:27:42.552798 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fnchd" podUID="bb732274-6b17-47f6-ace6-485706ecd538" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.041157 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.041883 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.041944 4979 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.042963 4979 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0"} pod="openshift-machine-config-operator/machine-config-daemon-8twmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.043034 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" containerID="cri-o://7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" gracePeriod=600 Dec 04 14:27:58 crc kubenswrapper[4979]: E1204 14:27:58.313570 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.386014 4979 generic.go:334] "Generic (PLEG): container finished" podID="d9b48d97-5715-467e-b044-4812d6dfea63" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" exitCode=0 Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.386082 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerDied","Data":"7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0"} Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.386519 4979 scope.go:117] "RemoveContainer" containerID="673e4ca9deca044cfcadec94d360ec56cdf6d25ae2da6226a912caa24ffed270" Dec 04 14:27:58 crc kubenswrapper[4979]: I1204 14:27:58.387725 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:27:58 crc kubenswrapper[4979]: E1204 14:27:58.388136 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.269468 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e9693784-a0d4-4fa6-8941-fd9f6617ae6c/init-config-reloader/0.log" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.506509 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e9693784-a0d4-4fa6-8941-fd9f6617ae6c/config-reloader/0.log" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.557210 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e9693784-a0d4-4fa6-8941-fd9f6617ae6c/init-config-reloader/0.log" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.561102 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_alertmanager-metric-storage-0_e9693784-a0d4-4fa6-8941-fd9f6617ae6c/alertmanager/0.log" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.841071 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_36d65388-9f95-4763-afb2-57c66bbc47e2/aodh-api/0.log" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.847313 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_36d65388-9f95-4763-afb2-57c66bbc47e2/aodh-evaluator/0.log" Dec 04 14:28:03 crc kubenswrapper[4979]: I1204 14:28:03.921754 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_36d65388-9f95-4763-afb2-57c66bbc47e2/aodh-listener/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.117892 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_36d65388-9f95-4763-afb2-57c66bbc47e2/aodh-notifier/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.174199 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5c694d4454-2bzkv_a014ea0c-872a-4cd9-a428-596d863e2651/barbican-api-log/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.253110 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5c694d4454-2bzkv_a014ea0c-872a-4cd9-a428-596d863e2651/barbican-api/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.411289 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68445fcf8b-jnsbd_3d82b744-e656-40b6-bb9c-c7ffc8c007cd/barbican-keystone-listener/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.492101 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-68445fcf8b-jnsbd_3d82b744-e656-40b6-bb9c-c7ffc8c007cd/barbican-keystone-listener-log/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.730899 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6bb568464f-gwp45_9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3/barbican-worker/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.736531 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6bb568464f-gwp45_9ffcbe68-bdb8-4a68-a99f-fb4d2adc0bd3/barbican-worker-log/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.888659 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-openstack-openstack-cell1-4f44q_902a1159-bb56-419e-ac6f-9cb05257d54f/bootstrap-openstack-openstack-cell1/0.log" Dec 04 14:28:04 crc kubenswrapper[4979]: I1204 14:28:04.984399 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_462e9872-5a8f-4fd0-bd81-cf2ec4471b60/ceilometer-central-agent/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.113636 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_462e9872-5a8f-4fd0-bd81-cf2ec4471b60/ceilometer-notification-agent/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.162583 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_462e9872-5a8f-4fd0-bd81-cf2ec4471b60/proxy-httpd/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.271609 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_462e9872-5a8f-4fd0-bd81-cf2ec4471b60/sg-core/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.375291 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-openstack-openstack-cell1-b5xn4_f1e47e68-9d7a-4202-97ff-eab1b0e2ccba/ceph-client-openstack-openstack-cell1/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.650547 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_96fc371f-833c-45a0-81c6-b91bfb6cb1c0/cinder-api/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.726861 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_96fc371f-833c-45a0-81c6-b91bfb6cb1c0/cinder-api-log/0.log" Dec 04 14:28:05 crc kubenswrapper[4979]: I1204 14:28:05.982897 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c98898f0-b483-4f47-9b32-101da166e4b7/probe/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.059941 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_c98898f0-b483-4f47-9b32-101da166e4b7/cinder-backup/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.354536 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3609683f-00ae-4919-8220-38c41870c261/cinder-scheduler/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.406924 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_3609683f-00ae-4919-8220-38c41870c261/probe/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.609142 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a/probe/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.718871 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_a76a87b5-d6c7-4b1a-92b1-6aad4febdc9a/cinder-volume/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.750875 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-openstack-openstack-cell1-rfcs6_e81e9d5c-a38e-4340-aca1-548c5b31f652/configure-network-openstack-openstack-cell1/0.log" Dec 04 14:28:06 crc kubenswrapper[4979]: I1204 14:28:06.936342 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-openstack-openstack-cell1-z8bpb_35f1f69f-68fd-4931-84ae-52547e563860/configure-os-openstack-openstack-cell1/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.003559 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5498899cc5-dshwr_7574d1f6-997e-4d18-8d8a-c49ba3b0f12e/init/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.204637 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5498899cc5-dshwr_7574d1f6-997e-4d18-8d8a-c49ba3b0f12e/init/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.226882 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5498899cc5-dshwr_7574d1f6-997e-4d18-8d8a-c49ba3b0f12e/dnsmasq-dns/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.263115 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-openstack-openstack-cell1-6dj6r_78c8481a-65fe-4e8a-85fc-d55cfb0dd709/download-cache-openstack-openstack-cell1/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.486825 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b4a92f9a-86bc-45a1-9172-2fbc0e9be453/glance-httpd/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.493883 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_b4a92f9a-86bc-45a1-9172-2fbc0e9be453/glance-log/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.597399 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_838f0563-3ad1-4f68-9727-40140a0ba82a/glance-httpd/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.658986 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_838f0563-3ad1-4f68-9727-40140a0ba82a/glance-log/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.886672 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-77ff7f9667-pk7gp_28eff767-0937-4b93-950d-d35c099e3ce5/heat-api/0.log" Dec 04 14:28:07 crc kubenswrapper[4979]: I1204 14:28:07.969814 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-68b7f67857-gnjfl_074b9676-de98-460f-9979-2790fc9778f5/heat-cfnapi/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.139228 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-7fd6595c6b-wn2bx_2628d26a-412b-43da-8525-b5375ea17e83/heat-engine/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.338046 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-788b9df87f-r5rlr_a43dcb59-d096-43aa-8e38-465be32822e8/horizon-log/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.373853 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-788b9df87f-r5rlr_a43dcb59-d096-43aa-8e38-465be32822e8/horizon/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.418137 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-openstack-openstack-cell1-vlmvv_2fc0f4a7-a63e-416b-b4da-b5f488c06289/install-certs-openstack-openstack-cell1/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.578553 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-openstack-openstack-cell1-jdvdr_7fed6f42-de65-4d92-9532-57de104f61de/install-os-openstack-openstack-cell1/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.764704 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-55d998f8fd-5tnfq_783f6a9e-4f95-4180-948e-122ec32a0deb/keystone-api/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.805675 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29414281-drh6b_55f01a95-3d30-4d74-8cf7-e1b5b35cb6a6/keystone-cron/0.log" Dec 04 14:28:08 crc kubenswrapper[4979]: I1204 14:28:08.885955 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_116e4048-6b4c-42b7-a6d8-f0b436b1b176/kube-state-metrics/0.log" Dec 04 14:28:09 crc kubenswrapper[4979]: I1204 14:28:09.027268 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-openstack-openstack-cell1-62gcp_6e6e1861-f679-4685-9e3a-9bf6b9fb00a8/libvirt-openstack-openstack-cell1/0.log" Dec 04 14:28:09 crc kubenswrapper[4979]: I1204 14:28:09.172745 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_bbb3740a-8c4b-4c8b-9248-552fcb198bf7/manila-api-log/0.log" Dec 04 14:28:09 crc kubenswrapper[4979]: I1204 14:28:09.223440 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_bbb3740a-8c4b-4c8b-9248-552fcb198bf7/manila-api/0.log" Dec 04 14:28:09 crc kubenswrapper[4979]: I1204 14:28:09.366471 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_defcd96a-640e-44a5-881e-b45711312d7e/probe/0.log" Dec 04 14:28:09 crc kubenswrapper[4979]: I1204 14:28:09.397858 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_defcd96a-640e-44a5-881e-b45711312d7e/manila-scheduler/0.log" Dec 04 14:28:10 crc kubenswrapper[4979]: I1204 14:28:10.163993 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_590f125d-d9fc-4847-a4b8-6f5619a6fbcd/probe/0.log" Dec 04 14:28:10 crc kubenswrapper[4979]: I1204 14:28:10.204355 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:28:10 crc kubenswrapper[4979]: E1204 14:28:10.204735 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:28:10 crc kubenswrapper[4979]: I1204 14:28:10.209834 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_590f125d-d9fc-4847-a4b8-6f5619a6fbcd/manila-share/0.log" Dec 04 14:28:10 crc kubenswrapper[4979]: I1204 14:28:10.821363 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5968f5544f-zdvnf_3e7547b4-b2b6-4917-b872-1282d7a8a876/neutron-httpd/0.log" Dec 04 14:28:10 crc kubenswrapper[4979]: I1204 14:28:10.845116 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5968f5544f-zdvnf_3e7547b4-b2b6-4917-b872-1282d7a8a876/neutron-api/0.log" Dec 04 14:28:11 crc kubenswrapper[4979]: I1204 14:28:11.037034 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-dhcp-openstack-openstack-cell1-2rrr6_7b19b05f-caa7-45f7-9eb5-890ef8fecaae/neutron-dhcp-openstack-openstack-cell1/0.log" Dec 04 14:28:11 crc kubenswrapper[4979]: I1204 14:28:11.171339 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-openstack-openstack-cell1-p86wx_41360a09-d209-4e16-9268-791d9b9afa82/neutron-metadata-openstack-openstack-cell1/0.log" Dec 04 14:28:11 crc kubenswrapper[4979]: I1204 14:28:11.887768 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-sriov-openstack-openstack-cell1-h9f9k_d396b2b7-0262-4e9d-a170-b7cb958b2c3c/neutron-sriov-openstack-openstack-cell1/0.log" Dec 04 14:28:12 crc kubenswrapper[4979]: I1204 14:28:12.005812 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c6124ed0-efd3-4275-982f-994549126581/nova-api-api/0.log" Dec 04 14:28:12 crc kubenswrapper[4979]: I1204 14:28:12.060758 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_c6124ed0-efd3-4275-982f-994549126581/nova-api-log/0.log" Dec 04 14:28:12 crc kubenswrapper[4979]: I1204 14:28:12.228723 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_1e0ffe92-c44a-454c-838f-d53f10f4a5f4/nova-cell0-conductor-conductor/0.log" Dec 04 14:28:12 crc kubenswrapper[4979]: I1204 14:28:12.492032 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_d082d430-346e-43ac-adc2-cfce3ff1d996/nova-cell1-conductor-conductor/0.log" Dec 04 14:28:12 crc kubenswrapper[4979]: I1204 14:28:12.624220 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_b0826aec-f059-4463-9942-3c84bd8a556d/nova-cell1-novncproxy-novncproxy/0.log" Dec 04 14:28:12 crc kubenswrapper[4979]: I1204 14:28:12.870961 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cellzgrts_351d7f45-a24f-46da-a9c9-2b598fff069d/nova-cell1-openstack-nova-compute-ffu-cell1-openstack-cell1/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.036877 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-openstack-openstack-cell1-8cfz2_3250581b-c62c-4c80-9738-c997e7600719/nova-cell1-openstack-openstack-cell1/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.124170 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b3cd38ce-827c-43bd-84ad-c676c62c8eaf/nova-metadata-log/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.236723 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b3cd38ce-827c-43bd-84ad-c676c62c8eaf/nova-metadata-metadata/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.458025 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_55b8d9b5-db3e-41fa-b712-02f19a7f3bfe/nova-scheduler-scheduler/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.493688 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7d99f6b98b-p9sbq_e3146a9d-fd80-475d-a9c2-cf52b42b437a/init/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.731673 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7d99f6b98b-p9sbq_e3146a9d-fd80-475d-a9c2-cf52b42b437a/init/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.852952 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7d99f6b98b-p9sbq_e3146a9d-fd80-475d-a9c2-cf52b42b437a/octavia-api-provider-agent/0.log" Dec 04 14:28:13 crc kubenswrapper[4979]: I1204 14:28:13.981800 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-ktcdr_cf209239-4f9e-43dd-b5e0-fc1253c4fd7a/init/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.114991 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-api-7d99f6b98b-p9sbq_e3146a9d-fd80-475d-a9c2-cf52b42b437a/octavia-api/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.251660 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-ktcdr_cf209239-4f9e-43dd-b5e0-fc1253c4fd7a/init/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.321961 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-healthmanager-ktcdr_cf209239-4f9e-43dd-b5e0-fc1253c4fd7a/octavia-healthmanager/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.421952 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-fmc22_f724c253-8e48-4d2c-ae4a-f24351b450a0/init/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.594428 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-fmc22_f724c253-8e48-4d2c-ae4a-f24351b450a0/octavia-housekeeping/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.596636 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-housekeeping-fmc22_f724c253-8e48-4d2c-ae4a-f24351b450a0/init/0.log" Dec 04 14:28:14 crc kubenswrapper[4979]: I1204 14:28:14.709757 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-ghhlt_abea20e4-7ff1-4f96-abcd-8c0a86c7df7e/init/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.119859 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-ghhlt_abea20e4-7ff1-4f96-abcd-8c0a86c7df7e/octavia-amphora-httpd/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.196663 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-image-upload-59f8cff499-ghhlt_abea20e4-7ff1-4f96-abcd-8c0a86c7df7e/init/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.242621 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-5bm9b_1d361aab-dd58-4fc9-aa52-738ff495f640/init/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.416489 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-5bm9b_1d361aab-dd58-4fc9-aa52-738ff495f640/octavia-rsyslog/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.440000 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-rsyslog-5bm9b_1d361aab-dd58-4fc9-aa52-738ff495f640/init/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.507468 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-w2lng_6efdadb0-0d97-4a27-9fa4-df4610542d61/init/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.808503 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-w2lng_6efdadb0-0d97-4a27-9fa4-df4610542d61/init/0.log" Dec 04 14:28:15 crc kubenswrapper[4979]: I1204 14:28:15.889895 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f6a16c64-130c-4c9a-988b-a3fa304c4d1e/mysql-bootstrap/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.121190 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_octavia-worker-w2lng_6efdadb0-0d97-4a27-9fa4-df4610542d61/octavia-worker/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.136809 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f6a16c64-130c-4c9a-988b-a3fa304c4d1e/mysql-bootstrap/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.146277 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_f6a16c64-130c-4c9a-988b-a3fa304c4d1e/galera/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.393336 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a4f9a655-3c5c-4053-bf72-1b99d3631889/mysql-bootstrap/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.675813 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a4f9a655-3c5c-4053-bf72-1b99d3631889/mysql-bootstrap/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.735182 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_a4f9a655-3c5c-4053-bf72-1b99d3631889/galera/0.log" Dec 04 14:28:16 crc kubenswrapper[4979]: I1204 14:28:16.784079 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_e55a2ea6-d990-412d-b5fb-21631921a849/openstackclient/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.030186 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-cv4bf_75e9db6b-7c43-497c-a993-34ea88e91b8f/ovn-controller/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.162632 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-4kwz7_cd2b947c-fa08-4f71-88ed-cd2cc216cc1e/openstack-network-exporter/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.350835 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dfk2m_848fa47f-ef4c-4eef-9a0f-e0536131bba3/ovsdb-server-init/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.579025 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dfk2m_848fa47f-ef4c-4eef-9a0f-e0536131bba3/ovs-vswitchd/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.594755 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dfk2m_848fa47f-ef4c-4eef-9a0f-e0536131bba3/ovsdb-server-init/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.637360 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dfk2m_848fa47f-ef4c-4eef-9a0f-e0536131bba3/ovsdb-server/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.876663 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_0ab214c3-6f92-443f-81e7-c48d306eba45/openstack-network-exporter/0.log" Dec 04 14:28:17 crc kubenswrapper[4979]: I1204 14:28:17.914738 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_0ab214c3-6f92-443f-81e7-c48d306eba45/ovn-northd/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.104702 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-openstack-openstack-cell1-8npfx_bf790ff5-131c-4db5-b428-2b920b290579/ovn-openstack-openstack-cell1/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.269491 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_62c0f621-2918-4193-9420-780b3ecbd5e5/openstack-network-exporter/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.299644 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_62c0f621-2918-4193-9420-780b3ecbd5e5/ovsdbserver-nb/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.476313 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_2e0f5920-9f22-4288-b09f-be3e565d88df/openstack-network-exporter/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.524742 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-1_2e0f5920-9f22-4288-b09f-be3e565d88df/ovsdbserver-nb/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.684557 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_d3ac5ba2-798d-4101-ad63-14cf9cfd47b4/openstack-network-exporter/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.748859 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-2_d3ac5ba2-798d-4101-ad63-14cf9cfd47b4/ovsdbserver-nb/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.895118 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6daaa5ce-1809-4b88-bec5-90204cbc1491/openstack-network-exporter/0.log" Dec 04 14:28:18 crc kubenswrapper[4979]: I1204 14:28:18.986473 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_6daaa5ce-1809-4b88-bec5-90204cbc1491/ovsdbserver-sb/0.log" Dec 04 14:28:19 crc kubenswrapper[4979]: I1204 14:28:19.097722 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_4d9841cd-dc04-4f50-8436-3d020df5f2f2/openstack-network-exporter/0.log" Dec 04 14:28:19 crc kubenswrapper[4979]: I1204 14:28:19.179402 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-1_4d9841cd-dc04-4f50-8436-3d020df5f2f2/ovsdbserver-sb/0.log" Dec 04 14:28:19 crc kubenswrapper[4979]: I1204 14:28:19.345031 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_dd92753a-87a9-47ca-ae6b-add04cbf46d3/openstack-network-exporter/0.log" Dec 04 14:28:19 crc kubenswrapper[4979]: I1204 14:28:19.550623 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-2_dd92753a-87a9-47ca-ae6b-add04cbf46d3/ovsdbserver-sb/0.log" Dec 04 14:28:19 crc kubenswrapper[4979]: I1204 14:28:19.933871 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69d694d89b-b86c2_833d82eb-d525-4813-8ad4-6ba2209832af/placement-api/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.058099 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-69d694d89b-b86c2_833d82eb-d525-4813-8ad4-6ba2209832af/placement-log/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.107568 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_pre-adoption-validation-openstack-pre-adoption-openstack-cvj8cf_39f115b5-6b82-4b63-848e-3cc6ae15099e/pre-adoption-validation-openstack-pre-adoption-openstack-cell1/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.186487 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_116becdc-fe97-4de3-902f-77dfd1332c5b/init-config-reloader/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.443066 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_116becdc-fe97-4de3-902f-77dfd1332c5b/init-config-reloader/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.505016 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_116becdc-fe97-4de3-902f-77dfd1332c5b/thanos-sidecar/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.529686 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_116becdc-fe97-4de3-902f-77dfd1332c5b/config-reloader/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.582085 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_116becdc-fe97-4de3-902f-77dfd1332c5b/prometheus/0.log" Dec 04 14:28:20 crc kubenswrapper[4979]: I1204 14:28:20.697536 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_61f5d1b2-dc4a-49f5-8c2c-ce7193287348/setup-container/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.151369 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_61f5d1b2-dc4a-49f5-8c2c-ce7193287348/setup-container/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.170146 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2149bf59-192d-42c6-98a8-989643a5094b/setup-container/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.187362 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_61f5d1b2-dc4a-49f5-8c2c-ce7193287348/rabbitmq/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.199336 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:28:21 crc kubenswrapper[4979]: E1204 14:28:21.199543 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.426151 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2149bf59-192d-42c6-98a8-989643a5094b/setup-container/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.482956 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_2149bf59-192d-42c6-98a8-989643a5094b/rabbitmq/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.558972 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-openstack-openstack-cell1-r6nzl_ff6b5d75-375e-4c64-a827-089952fb1d92/reboot-os-openstack-openstack-cell1/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.707946 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-openstack-openstack-cell1-dncws_c75e0fbc-b974-4306-83b7-20c82ac841e7/run-os-openstack-openstack-cell1/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.841709 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_af82841c-53fb-4e62-8fd3-fbf888e87045/memcached/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.841879 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-openstack-5rvvt_63ae0b49-5321-4eff-96c5-3c3e22a1ea8d/ssh-known-hosts-openstack/0.log" Dec 04 14:28:21 crc kubenswrapper[4979]: I1204 14:28:21.995085 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-openstack-openstack-cell1-szvj8_3aacedd4-db12-48a8-9ad0-1c95d92b24f4/telemetry-openstack-openstack-cell1/0.log" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.204006 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tripleo-cleanup-tripleo-cleanup-openstack-cell1-lxgr7_717bd66f-9118-4b81-927c-27f66aa6483a/tripleo-cleanup-tripleo-cleanup-openstack-cell1/0.log" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.250643 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-openstack-openstack-cell1-nq7vm_8647a121-d252-4072-aded-c1166beefd89/validate-network-openstack-openstack-cell1/0.log" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.944670 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-g5fn9"] Dec 04 14:28:22 crc kubenswrapper[4979]: E1204 14:28:22.945250 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="extract-utilities" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945274 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="extract-utilities" Dec 04 14:28:22 crc kubenswrapper[4979]: E1204 14:28:22.945314 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="registry-server" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945324 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="registry-server" Dec 04 14:28:22 crc kubenswrapper[4979]: E1204 14:28:22.945334 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="extract-content" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945342 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="extract-content" Dec 04 14:28:22 crc kubenswrapper[4979]: E1204 14:28:22.945373 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="registry-server" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945381 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="registry-server" Dec 04 14:28:22 crc kubenswrapper[4979]: E1204 14:28:22.945394 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="extract-content" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945401 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="extract-content" Dec 04 14:28:22 crc kubenswrapper[4979]: E1204 14:28:22.945415 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="extract-utilities" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945422 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="extract-utilities" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945861 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="74114e27-f36b-47b2-8088-cae23a80ddae" containerName="registry-server" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.945892 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="006b9d3b-f389-46ae-a38e-ed08e9b42fce" containerName="registry-server" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.950455 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:22 crc kubenswrapper[4979]: I1204 14:28:22.983527 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5fn9"] Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.045657 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-catalog-content\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.045785 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cxrq\" (UniqueName: \"kubernetes.io/projected/42618389-b8fc-43ce-9830-cd37e53ee691-kube-api-access-2cxrq\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.045822 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-utilities\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.148000 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-catalog-content\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.148092 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cxrq\" (UniqueName: \"kubernetes.io/projected/42618389-b8fc-43ce-9830-cd37e53ee691-kube-api-access-2cxrq\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.148119 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-utilities\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.148711 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-catalog-content\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.148795 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-utilities\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.171610 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cxrq\" (UniqueName: \"kubernetes.io/projected/42618389-b8fc-43ce-9830-cd37e53ee691-kube-api-access-2cxrq\") pod \"redhat-marketplace-g5fn9\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.288595 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:23 crc kubenswrapper[4979]: I1204 14:28:23.803204 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5fn9"] Dec 04 14:28:24 crc kubenswrapper[4979]: I1204 14:28:24.713683 4979 generic.go:334] "Generic (PLEG): container finished" podID="42618389-b8fc-43ce-9830-cd37e53ee691" containerID="ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3" exitCode=0 Dec 04 14:28:24 crc kubenswrapper[4979]: I1204 14:28:24.714137 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerDied","Data":"ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3"} Dec 04 14:28:24 crc kubenswrapper[4979]: I1204 14:28:24.714164 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerStarted","Data":"b91d21f41eb13c07193fc326fcf83dcc0b36a37965d51a0958a89bf4a2e82fd5"} Dec 04 14:28:25 crc kubenswrapper[4979]: I1204 14:28:25.725745 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerStarted","Data":"c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd"} Dec 04 14:28:26 crc kubenswrapper[4979]: I1204 14:28:26.737281 4979 generic.go:334] "Generic (PLEG): container finished" podID="42618389-b8fc-43ce-9830-cd37e53ee691" containerID="c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd" exitCode=0 Dec 04 14:28:26 crc kubenswrapper[4979]: I1204 14:28:26.737356 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerDied","Data":"c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd"} Dec 04 14:28:27 crc kubenswrapper[4979]: I1204 14:28:27.755670 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerStarted","Data":"9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b"} Dec 04 14:28:27 crc kubenswrapper[4979]: I1204 14:28:27.789401 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-g5fn9" podStartSLOduration=3.364171745 podStartE2EDuration="5.789378294s" podCreationTimestamp="2025-12-04 14:28:22 +0000 UTC" firstStartedPulling="2025-12-04 14:28:24.719397862 +0000 UTC m=+9928.993693666" lastFinishedPulling="2025-12-04 14:28:27.144604411 +0000 UTC m=+9931.418900215" observedRunningTime="2025-12-04 14:28:27.777454475 +0000 UTC m=+9932.051750279" watchObservedRunningTime="2025-12-04 14:28:27.789378294 +0000 UTC m=+9932.063674128" Dec 04 14:28:33 crc kubenswrapper[4979]: I1204 14:28:33.289215 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:33 crc kubenswrapper[4979]: I1204 14:28:33.289880 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:33 crc kubenswrapper[4979]: I1204 14:28:33.341198 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:33 crc kubenswrapper[4979]: I1204 14:28:33.864976 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:33 crc kubenswrapper[4979]: I1204 14:28:33.921050 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5fn9"] Dec 04 14:28:35 crc kubenswrapper[4979]: I1204 14:28:35.836977 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-g5fn9" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="registry-server" containerID="cri-o://9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b" gracePeriod=2 Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.251894 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:28:36 crc kubenswrapper[4979]: E1204 14:28:36.252721 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.596554 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.744230 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cxrq\" (UniqueName: \"kubernetes.io/projected/42618389-b8fc-43ce-9830-cd37e53ee691-kube-api-access-2cxrq\") pod \"42618389-b8fc-43ce-9830-cd37e53ee691\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.744520 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-utilities\") pod \"42618389-b8fc-43ce-9830-cd37e53ee691\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.744593 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-catalog-content\") pod \"42618389-b8fc-43ce-9830-cd37e53ee691\" (UID: \"42618389-b8fc-43ce-9830-cd37e53ee691\") " Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.745353 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-utilities" (OuterVolumeSpecName: "utilities") pod "42618389-b8fc-43ce-9830-cd37e53ee691" (UID: "42618389-b8fc-43ce-9830-cd37e53ee691"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.753481 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42618389-b8fc-43ce-9830-cd37e53ee691-kube-api-access-2cxrq" (OuterVolumeSpecName: "kube-api-access-2cxrq") pod "42618389-b8fc-43ce-9830-cd37e53ee691" (UID: "42618389-b8fc-43ce-9830-cd37e53ee691"). InnerVolumeSpecName "kube-api-access-2cxrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.762260 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42618389-b8fc-43ce-9830-cd37e53ee691" (UID: "42618389-b8fc-43ce-9830-cd37e53ee691"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.847404 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.847702 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42618389-b8fc-43ce-9830-cd37e53ee691-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.847714 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cxrq\" (UniqueName: \"kubernetes.io/projected/42618389-b8fc-43ce-9830-cd37e53ee691-kube-api-access-2cxrq\") on node \"crc\" DevicePath \"\"" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.851096 4979 generic.go:334] "Generic (PLEG): container finished" podID="42618389-b8fc-43ce-9830-cd37e53ee691" containerID="9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b" exitCode=0 Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.851140 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerDied","Data":"9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b"} Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.851168 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-g5fn9" event={"ID":"42618389-b8fc-43ce-9830-cd37e53ee691","Type":"ContainerDied","Data":"b91d21f41eb13c07193fc326fcf83dcc0b36a37965d51a0958a89bf4a2e82fd5"} Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.851184 4979 scope.go:117] "RemoveContainer" containerID="9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.851337 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-g5fn9" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.884459 4979 scope.go:117] "RemoveContainer" containerID="c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.890565 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5fn9"] Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.907979 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-g5fn9"] Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.908536 4979 scope.go:117] "RemoveContainer" containerID="ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.963590 4979 scope.go:117] "RemoveContainer" containerID="9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b" Dec 04 14:28:36 crc kubenswrapper[4979]: E1204 14:28:36.964076 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b\": container with ID starting with 9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b not found: ID does not exist" containerID="9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.964144 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b"} err="failed to get container status \"9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b\": rpc error: code = NotFound desc = could not find container \"9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b\": container with ID starting with 9e0f3034d2bb8fba19dc178dc2210d5498395b33e55c91e9e61f1d916560bc2b not found: ID does not exist" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.964174 4979 scope.go:117] "RemoveContainer" containerID="c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd" Dec 04 14:28:36 crc kubenswrapper[4979]: E1204 14:28:36.964498 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd\": container with ID starting with c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd not found: ID does not exist" containerID="c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.964542 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd"} err="failed to get container status \"c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd\": rpc error: code = NotFound desc = could not find container \"c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd\": container with ID starting with c686b5c056219ea615404e362cf3517c1649c467a0d73ced5209d37e654914cd not found: ID does not exist" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.964568 4979 scope.go:117] "RemoveContainer" containerID="ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3" Dec 04 14:28:36 crc kubenswrapper[4979]: E1204 14:28:36.964799 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3\": container with ID starting with ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3 not found: ID does not exist" containerID="ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3" Dec 04 14:28:36 crc kubenswrapper[4979]: I1204 14:28:36.964829 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3"} err="failed to get container status \"ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3\": rpc error: code = NotFound desc = could not find container \"ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3\": container with ID starting with ffc00ab33076f8ec0e230014513fa7609cb9a490705f65da0924bd8450db0bb3 not found: ID does not exist" Dec 04 14:28:38 crc kubenswrapper[4979]: I1204 14:28:38.247086 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" path="/var/lib/kubelet/pods/42618389-b8fc-43ce-9830-cd37e53ee691/volumes" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.267048 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/util/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.400509 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/util/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.435474 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/pull/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.502280 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/pull/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.681465 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/util/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.744056 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/pull/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.747876 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_7b975d6399606046f1762845d19d1c0dc235fd1f75e224180ebc3c0ac6jq5qm_6c9ef7f7-e93d-4c20-b799-9531bc5627a8/extract/0.log" Dec 04 14:28:47 crc kubenswrapper[4979]: I1204 14:28:47.927516 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5vbf8_cc4f8bbc-9399-4c17-a7a9-ec39a429813c/kube-rbac-proxy/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.058237 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-lw7hk_aa22fecb-5237-48bd-9526-857132083b7f/kube-rbac-proxy/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.066694 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-5vbf8_cc4f8bbc-9399-4c17-a7a9-ec39a429813c/manager/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.263713 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-lw7hk_aa22fecb-5237-48bd-9526-857132083b7f/manager/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.392334 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-kw2fk_3369e6a8-2d6d-4c72-b96a-2cf650c4c257/kube-rbac-proxy/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.395471 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-kw2fk_3369e6a8-2d6d-4c72-b96a-2cf650c4c257/manager/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.608164 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-mr2rj_5190c637-8327-4a8b-b351-4a6e86f6d59a/kube-rbac-proxy/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.770480 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-hgslt_01d00bed-f29a-4ada-98d1-18ac5935eb06/kube-rbac-proxy/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.820404 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-mr2rj_5190c637-8327-4a8b-b351-4a6e86f6d59a/manager/0.log" Dec 04 14:28:48 crc kubenswrapper[4979]: I1204 14:28:48.882869 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-hgslt_01d00bed-f29a-4ada-98d1-18ac5935eb06/manager/0.log" Dec 04 14:28:49 crc kubenswrapper[4979]: I1204 14:28:49.037283 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5rcqm_bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6/kube-rbac-proxy/0.log" Dec 04 14:28:49 crc kubenswrapper[4979]: I1204 14:28:49.417545 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-2pnbm_11b2439d-9796-4f42-94b4-f0903aa555d6/kube-rbac-proxy/0.log" Dec 04 14:28:49 crc kubenswrapper[4979]: I1204 14:28:49.445722 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5rcqm_bd63d5e5-4cb3-4749-8f86-88a32e3ba6a6/manager/0.log" Dec 04 14:28:49 crc kubenswrapper[4979]: I1204 14:28:49.757783 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7gb6r_316f3ead-f9a1-44c5-84f9-5d8304a9f15b/kube-rbac-proxy/0.log" Dec 04 14:28:49 crc kubenswrapper[4979]: I1204 14:28:49.798441 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7gb6r_316f3ead-f9a1-44c5-84f9-5d8304a9f15b/manager/0.log" Dec 04 14:28:49 crc kubenswrapper[4979]: I1204 14:28:49.851273 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-2pnbm_11b2439d-9796-4f42-94b4-f0903aa555d6/manager/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.091873 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-q2kv7_0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b/kube-rbac-proxy/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.185984 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-q2kv7_0f7b9cfe-b0ea-4062-81b2-bc6d94435b3b/manager/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.199403 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:28:50 crc kubenswrapper[4979]: E1204 14:28:50.199697 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.256908 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-hwjs2_fe9dded8-cafd-469a-b65d-cfa21673e4b1/kube-rbac-proxy/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.352229 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-hwjs2_fe9dded8-cafd-469a-b65d-cfa21673e4b1/manager/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.541162 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-tv9lc_465158ac-c7e8-4df8-97ee-231ba9ca3541/manager/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.585881 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-tv9lc_465158ac-c7e8-4df8-97ee-231ba9ca3541/kube-rbac-proxy/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.675213 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-xzkn7_d7efdf6b-7859-4750-b531-8bae5f2332fd/kube-rbac-proxy/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.906049 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-xzkn7_d7efdf6b-7859-4750-b531-8bae5f2332fd/manager/0.log" Dec 04 14:28:50 crc kubenswrapper[4979]: I1204 14:28:50.920091 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-xwtfq_4d6d4d26-72f0-4b63-979c-a533eb0dd53f/kube-rbac-proxy/0.log" Dec 04 14:28:51 crc kubenswrapper[4979]: I1204 14:28:51.272672 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-xwtfq_4d6d4d26-72f0-4b63-979c-a533eb0dd53f/manager/0.log" Dec 04 14:28:51 crc kubenswrapper[4979]: I1204 14:28:51.293964 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-4j4cc_e729014a-4926-4219-9159-785c4f602117/kube-rbac-proxy/0.log" Dec 04 14:28:51 crc kubenswrapper[4979]: I1204 14:28:51.313369 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-4j4cc_e729014a-4926-4219-9159-785c4f602117/manager/0.log" Dec 04 14:28:51 crc kubenswrapper[4979]: I1204 14:28:51.518161 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4xr262_3583aaeb-1aeb-4afc-91d2-a6837635a371/manager/0.log" Dec 04 14:28:51 crc kubenswrapper[4979]: I1204 14:28:51.526897 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4xr262_3583aaeb-1aeb-4afc-91d2-a6837635a371/kube-rbac-proxy/0.log" Dec 04 14:28:51 crc kubenswrapper[4979]: I1204 14:28:51.975506 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xngxq_26a7a95d-a941-480c-b4a8-d2e892c26a45/registry-server/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.079136 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-56668d8dc4-9dsck_c8d63092-cd3d-497c-87ea-4be3a31aff53/operator/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.242095 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xgjxw_7c730311-2e6a-4815-87f4-f8099d5ced63/kube-rbac-proxy/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.394095 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xgjxw_7c730311-2e6a-4815-87f4-f8099d5ced63/manager/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.446545 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7jr2r_30066915-9744-4901-82c8-d19d06b1fb49/kube-rbac-proxy/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.631117 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-7jr2r_30066915-9744-4901-82c8-d19d06b1fb49/manager/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.831653 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-kk4p2_e5a502b0-c2bf-43fb-87eb-8def7494a0b9/operator/0.log" Dec 04 14:28:52 crc kubenswrapper[4979]: I1204 14:28:52.930981 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-dhmqs_d8e79a30-90c1-409c-a12b-ad2d0e0d36c8/kube-rbac-proxy/0.log" Dec 04 14:28:53 crc kubenswrapper[4979]: I1204 14:28:53.079135 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-dhmqs_d8e79a30-90c1-409c-a12b-ad2d0e0d36c8/manager/0.log" Dec 04 14:28:53 crc kubenswrapper[4979]: I1204 14:28:53.331368 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-55g6q_91e3f7ac-b8d9-40d8-8a1c-94412c0a9488/kube-rbac-proxy/0.log" Dec 04 14:28:53 crc kubenswrapper[4979]: I1204 14:28:53.504371 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-55g6q_91e3f7ac-b8d9-40d8-8a1c-94412c0a9488/manager/0.log" Dec 04 14:28:53 crc kubenswrapper[4979]: I1204 14:28:53.538970 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-c9c8v_fa56245e-7ded-4797-9fd9-bc84274407e2/kube-rbac-proxy/0.log" Dec 04 14:28:53 crc kubenswrapper[4979]: I1204 14:28:53.741921 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-c9c8v_fa56245e-7ded-4797-9fd9-bc84274407e2/manager/0.log" Dec 04 14:28:53 crc kubenswrapper[4979]: I1204 14:28:53.911690 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-ddkhk_4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04/manager/0.log" Dec 04 14:28:54 crc kubenswrapper[4979]: I1204 14:28:53.999972 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-ddkhk_4c3301d5-6fdd-4e3c-a4ba-f2c11322fd04/kube-rbac-proxy/0.log" Dec 04 14:28:54 crc kubenswrapper[4979]: I1204 14:28:54.334118 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-79cddf9bf6-xlr4l_8ef350ec-4178-410b-9e8e-9981810dc148/manager/0.log" Dec 04 14:29:02 crc kubenswrapper[4979]: I1204 14:29:02.199352 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:29:02 crc kubenswrapper[4979]: E1204 14:29:02.200366 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:29:14 crc kubenswrapper[4979]: I1204 14:29:14.199159 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:29:14 crc kubenswrapper[4979]: E1204 14:29:14.200504 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:29:15 crc kubenswrapper[4979]: I1204 14:29:15.704940 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-dkjdk_a27662f3-0c05-4afb-addc-ebcfdf5a6b95/control-plane-machine-set-operator/0.log" Dec 04 14:29:16 crc kubenswrapper[4979]: I1204 14:29:16.375626 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5bsgf_9f27cd34-9226-4968-a4e7-83788b460159/machine-api-operator/0.log" Dec 04 14:29:16 crc kubenswrapper[4979]: I1204 14:29:16.394560 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-5bsgf_9f27cd34-9226-4968-a4e7-83788b460159/kube-rbac-proxy/0.log" Dec 04 14:29:27 crc kubenswrapper[4979]: I1204 14:29:27.199488 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:29:27 crc kubenswrapper[4979]: E1204 14:29:27.200254 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:29:29 crc kubenswrapper[4979]: I1204 14:29:29.287371 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-86cb77c54b-4p5pm_ff130626-7c18-4528-869b-0191bc75cca3/cert-manager-controller/0.log" Dec 04 14:29:29 crc kubenswrapper[4979]: I1204 14:29:29.440067 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-855d9ccff4-svrq5_af5b7f91-c64a-4f1b-8a48-3863730942c3/cert-manager-cainjector/0.log" Dec 04 14:29:29 crc kubenswrapper[4979]: I1204 14:29:29.470503 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-f4fb5df64-kqcwm_b762418e-2e6a-4678-ad5f-7acac1219d5e/cert-manager-webhook/0.log" Dec 04 14:29:41 crc kubenswrapper[4979]: I1204 14:29:41.199258 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:29:41 crc kubenswrapper[4979]: E1204 14:29:41.200035 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:29:41 crc kubenswrapper[4979]: I1204 14:29:41.666184 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5qhmm_ac09b8ec-7526-45ed-9d0a-56c8416c635a/nmstate-console-plugin/0.log" Dec 04 14:29:41 crc kubenswrapper[4979]: I1204 14:29:41.851479 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-fhws5_515dbf63-c417-4af6-b570-a58812752c7f/nmstate-handler/0.log" Dec 04 14:29:41 crc kubenswrapper[4979]: I1204 14:29:41.898136 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hwtsn_6d6ff593-472b-4ae3-a015-a846b703daf8/kube-rbac-proxy/0.log" Dec 04 14:29:41 crc kubenswrapper[4979]: I1204 14:29:41.942751 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-hwtsn_6d6ff593-472b-4ae3-a015-a846b703daf8/nmstate-metrics/0.log" Dec 04 14:29:42 crc kubenswrapper[4979]: I1204 14:29:42.043082 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-2pbw8_42b69f76-cf35-4413-9e2a-12ae411b6d7d/nmstate-operator/0.log" Dec 04 14:29:42 crc kubenswrapper[4979]: I1204 14:29:42.156687 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-7g9qj_e8a8bc91-603d-4d91-a54d-ec5c38845cb8/nmstate-webhook/0.log" Dec 04 14:29:53 crc kubenswrapper[4979]: I1204 14:29:53.198927 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:29:53 crc kubenswrapper[4979]: E1204 14:29:53.199816 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:29:58 crc kubenswrapper[4979]: I1204 14:29:58.591242 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d4d2t_f90cd359-ebe3-4c34-a2c0-16bf8805b876/kube-rbac-proxy/0.log" Dec 04 14:29:58 crc kubenswrapper[4979]: I1204 14:29:58.905782 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-frr-files/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.006405 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-d4d2t_f90cd359-ebe3-4c34-a2c0-16bf8805b876/controller/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.065102 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-frr-files/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.091942 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-reloader/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.122694 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-metrics/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.173893 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-reloader/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.385033 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-metrics/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.385118 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-frr-files/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.397786 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-metrics/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.413167 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-reloader/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.563828 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-metrics/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.571705 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-frr-files/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.593290 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/cp-reloader/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.660244 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/controller/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.749179 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/frr-metrics/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.791169 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/kube-rbac-proxy/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.862430 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/kube-rbac-proxy-frr/0.log" Dec 04 14:29:59 crc kubenswrapper[4979]: I1204 14:29:59.957793 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/reloader/0.log" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.084895 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-cw6w6_0e84c35a-57f3-43e8-9d36-6e9750b85c9d/frr-k8s-webhook-server/0.log" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.187911 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f"] Dec 04 14:30:00 crc kubenswrapper[4979]: E1204 14:30:00.188434 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="extract-content" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.188452 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="extract-content" Dec 04 14:30:00 crc kubenswrapper[4979]: E1204 14:30:00.188470 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="extract-utilities" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.188477 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="extract-utilities" Dec 04 14:30:00 crc kubenswrapper[4979]: E1204 14:30:00.188502 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="registry-server" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.188508 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="registry-server" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.188707 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="42618389-b8fc-43ce-9830-cd37e53ee691" containerName="registry-server" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.189453 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.193959 4979 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.208266 4979 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.257337 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e40a601c-3b93-4693-8bfb-71a87ed29676-secret-volume\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.257505 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e40a601c-3b93-4693-8bfb-71a87ed29676-config-volume\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.257580 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zch\" (UniqueName: \"kubernetes.io/projected/e40a601c-3b93-4693-8bfb-71a87ed29676-kube-api-access-c6zch\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.262613 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f"] Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.359710 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e40a601c-3b93-4693-8bfb-71a87ed29676-secret-volume\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.359918 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e40a601c-3b93-4693-8bfb-71a87ed29676-config-volume\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.360000 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zch\" (UniqueName: \"kubernetes.io/projected/e40a601c-3b93-4693-8bfb-71a87ed29676-kube-api-access-c6zch\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.365324 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e40a601c-3b93-4693-8bfb-71a87ed29676-config-volume\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.383898 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e40a601c-3b93-4693-8bfb-71a87ed29676-secret-volume\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.393087 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zch\" (UniqueName: \"kubernetes.io/projected/e40a601c-3b93-4693-8bfb-71a87ed29676-kube-api-access-c6zch\") pod \"collect-profiles-29414310-cmd8f\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.491932 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7d59d98bc-t8mjh_06d63892-a58b-4921-9163-acdb6808729d/manager/0.log" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.546321 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:00 crc kubenswrapper[4979]: I1204 14:30:00.686672 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86498874f8-zlxh7_6e81d70c-96a3-4a8d-b817-534a20d99799/webhook-server/0.log" Dec 04 14:30:01 crc kubenswrapper[4979]: I1204 14:30:01.009968 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-msnnj_78803b74-9eb0-4721-91c5-39490f93248a/kube-rbac-proxy/0.log" Dec 04 14:30:01 crc kubenswrapper[4979]: I1204 14:30:01.120217 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f"] Dec 04 14:30:01 crc kubenswrapper[4979]: I1204 14:30:01.888385 4979 generic.go:334] "Generic (PLEG): container finished" podID="e40a601c-3b93-4693-8bfb-71a87ed29676" containerID="69d8e92d4469804512e955f0d29dff27665b1484451840c105098364820ced1a" exitCode=0 Dec 04 14:30:01 crc kubenswrapper[4979]: I1204 14:30:01.888459 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" event={"ID":"e40a601c-3b93-4693-8bfb-71a87ed29676","Type":"ContainerDied","Data":"69d8e92d4469804512e955f0d29dff27665b1484451840c105098364820ced1a"} Dec 04 14:30:01 crc kubenswrapper[4979]: I1204 14:30:01.888695 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" event={"ID":"e40a601c-3b93-4693-8bfb-71a87ed29676","Type":"ContainerStarted","Data":"fe42f7d246013526fd3e4d9c020b1d346c13825fc6aeee1191ef6ffc148258f1"} Dec 04 14:30:02 crc kubenswrapper[4979]: I1204 14:30:02.062192 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-msnnj_78803b74-9eb0-4721-91c5-39490f93248a/speaker/0.log" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.331029 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.469099 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e40a601c-3b93-4693-8bfb-71a87ed29676-secret-volume\") pod \"e40a601c-3b93-4693-8bfb-71a87ed29676\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.469168 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e40a601c-3b93-4693-8bfb-71a87ed29676-config-volume\") pod \"e40a601c-3b93-4693-8bfb-71a87ed29676\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.469205 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6zch\" (UniqueName: \"kubernetes.io/projected/e40a601c-3b93-4693-8bfb-71a87ed29676-kube-api-access-c6zch\") pod \"e40a601c-3b93-4693-8bfb-71a87ed29676\" (UID: \"e40a601c-3b93-4693-8bfb-71a87ed29676\") " Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.471288 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e40a601c-3b93-4693-8bfb-71a87ed29676-config-volume" (OuterVolumeSpecName: "config-volume") pod "e40a601c-3b93-4693-8bfb-71a87ed29676" (UID: "e40a601c-3b93-4693-8bfb-71a87ed29676"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.480512 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40a601c-3b93-4693-8bfb-71a87ed29676-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e40a601c-3b93-4693-8bfb-71a87ed29676" (UID: "e40a601c-3b93-4693-8bfb-71a87ed29676"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.481910 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e40a601c-3b93-4693-8bfb-71a87ed29676-kube-api-access-c6zch" (OuterVolumeSpecName: "kube-api-access-c6zch") pod "e40a601c-3b93-4693-8bfb-71a87ed29676" (UID: "e40a601c-3b93-4693-8bfb-71a87ed29676"). InnerVolumeSpecName "kube-api-access-c6zch". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.571500 4979 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e40a601c-3b93-4693-8bfb-71a87ed29676-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.571747 4979 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e40a601c-3b93-4693-8bfb-71a87ed29676-config-volume\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.571814 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6zch\" (UniqueName: \"kubernetes.io/projected/e40a601c-3b93-4693-8bfb-71a87ed29676-kube-api-access-c6zch\") on node \"crc\" DevicePath \"\"" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.680532 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-sbk2k_b7760815-ac81-45eb-8a1a-0b6afe53f006/frr/0.log" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.908058 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" event={"ID":"e40a601c-3b93-4693-8bfb-71a87ed29676","Type":"ContainerDied","Data":"fe42f7d246013526fd3e4d9c020b1d346c13825fc6aeee1191ef6ffc148258f1"} Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.908653 4979 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe42f7d246013526fd3e4d9c020b1d346c13825fc6aeee1191ef6ffc148258f1" Dec 04 14:30:03 crc kubenswrapper[4979]: I1204 14:30:03.908102 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29414310-cmd8f" Dec 04 14:30:04 crc kubenswrapper[4979]: I1204 14:30:04.419859 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb"] Dec 04 14:30:04 crc kubenswrapper[4979]: I1204 14:30:04.430365 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29414265-rrqzb"] Dec 04 14:30:06 crc kubenswrapper[4979]: I1204 14:30:06.212927 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd899ee9-8952-40e6-b5af-d62e28fe2670" path="/var/lib/kubelet/pods/fd899ee9-8952-40e6-b5af-d62e28fe2670/volumes" Dec 04 14:30:08 crc kubenswrapper[4979]: I1204 14:30:08.199006 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:30:08 crc kubenswrapper[4979]: E1204 14:30:08.199917 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:30:14 crc kubenswrapper[4979]: I1204 14:30:14.985676 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/util/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.166046 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/util/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.222851 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/pull/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.258468 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/pull/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.411953 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/util/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.422912 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/extract/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.460162 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931akzz6n_4dc4b0b6-7e83-4e24-a772-a9607d3c8d97/pull/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.643951 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/util/0.log" Dec 04 14:30:15 crc kubenswrapper[4979]: I1204 14:30:15.780712 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/util/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.103902 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/pull/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.104047 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/pull/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.104176 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/pull/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.112002 4979 scope.go:117] "RemoveContainer" containerID="cb94fb6a5d541da09f48ab189b4bfd79abac67e41b34e33033216ee23b0fa9dd" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.343329 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/extract/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.344641 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/util/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.384863 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fg7wff_7ee51840-874e-4d0a-ac2c-5749ac25adec/util/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.495651 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/util/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.557199 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/pull/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.574110 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/pull/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.748119 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/pull/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.748327 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/util/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.776054 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210q9wfs_89020d69-4269-45d7-a016-59f4f74e53a6/extract/0.log" Dec 04 14:30:16 crc kubenswrapper[4979]: I1204 14:30:16.934916 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/util/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.092031 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/pull/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.191042 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/util/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.195372 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/pull/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.322073 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/util/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.373701 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/pull/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.422295 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83rltc7_08c193f7-63b8-4a74-bfd3-29ac10d5afd6/extract/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.563835 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/extract-utilities/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.737954 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/extract-utilities/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.768724 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/extract-content/0.log" Dec 04 14:30:17 crc kubenswrapper[4979]: I1204 14:30:17.771633 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/extract-content/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.015079 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/extract-utilities/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.042168 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/extract-content/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.154194 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/extract-utilities/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.481149 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/extract-content/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.488708 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/extract-content/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.574528 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/extract-utilities/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.728474 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/extract-utilities/0.log" Dec 04 14:30:18 crc kubenswrapper[4979]: I1204 14:30:18.821741 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/extract-content/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.018605 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-mvpph_54723683-07ae-4574-9396-43fc24d977af/marketplace-operator/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.126811 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/extract-utilities/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.321765 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/extract-utilities/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.405627 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/extract-content/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.417863 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/extract-content/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.464633 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-r8q2n_028b7d1f-52bf-4bc2-8905-2c1d5035b3fc/registry-server/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.653415 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/extract-utilities/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.662057 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/extract-content/0.log" Dec 04 14:30:19 crc kubenswrapper[4979]: I1204 14:30:19.916970 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/extract-utilities/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.170065 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-nfw4p_3c39667b-75fa-4174-896d-2424f435fde8/registry-server/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.186354 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-4xlng_5cc37280-3e5a-41d6-ae14-0bd9fc40ae42/registry-server/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.220505 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/extract-utilities/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.230499 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/extract-content/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.235623 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/extract-content/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.390310 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/extract-utilities/0.log" Dec 04 14:30:20 crc kubenswrapper[4979]: I1204 14:30:20.428779 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/extract-content/0.log" Dec 04 14:30:21 crc kubenswrapper[4979]: I1204 14:30:21.198731 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:30:21 crc kubenswrapper[4979]: E1204 14:30:21.199183 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:30:21 crc kubenswrapper[4979]: I1204 14:30:21.590161 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-bdrvn_a6359145-7b66-40db-833e-4bb609833266/registry-server/0.log" Dec 04 14:30:32 crc kubenswrapper[4979]: I1204 14:30:32.199451 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:30:32 crc kubenswrapper[4979]: E1204 14:30:32.200283 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:30:34 crc kubenswrapper[4979]: I1204 14:30:34.845672 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-5wrx5_e2d039ea-6c89-4435-88ea-e5ef847fe97c/prometheus-operator/0.log" Dec 04 14:30:35 crc kubenswrapper[4979]: I1204 14:30:35.027670 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67c858fbfb-qbjxv_8a3ef5ba-d73a-42e2-9942-6ba8fc02da27/prometheus-operator-admission-webhook/0.log" Dec 04 14:30:35 crc kubenswrapper[4979]: I1204 14:30:35.053967 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-67c858fbfb-tq5rh_0e7abc19-0af1-4edc-89bd-4ec0ab996bac/prometheus-operator-admission-webhook/0.log" Dec 04 14:30:35 crc kubenswrapper[4979]: I1204 14:30:35.261654 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-n6rcx_eeec64e3-82f9-4370-8621-fa7963371a38/perses-operator/0.log" Dec 04 14:30:35 crc kubenswrapper[4979]: I1204 14:30:35.265705 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-clrnn_0264a7ca-8bc9-43e6-8995-885f46a759dc/operator/0.log" Dec 04 14:30:44 crc kubenswrapper[4979]: I1204 14:30:44.201421 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:30:44 crc kubenswrapper[4979]: E1204 14:30:44.202368 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:30:55 crc kubenswrapper[4979]: I1204 14:30:55.199517 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:30:55 crc kubenswrapper[4979]: E1204 14:30:55.200434 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:30:58 crc kubenswrapper[4979]: E1204 14:30:58.948663 4979 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.74:58226->38.102.83.74:36883: read tcp 38.102.83.74:58226->38.102.83.74:36883: read: connection reset by peer Dec 04 14:31:05 crc kubenswrapper[4979]: E1204 14:31:05.595421 4979 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.74:46584->38.102.83.74:36883: write tcp 38.102.83.74:46584->38.102.83.74:36883: write: broken pipe Dec 04 14:31:06 crc kubenswrapper[4979]: E1204 14:31:06.012618 4979 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.74:46678->38.102.83.74:36883: write tcp 38.102.83.74:46678->38.102.83.74:36883: write: broken pipe Dec 04 14:31:09 crc kubenswrapper[4979]: I1204 14:31:09.200371 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:31:09 crc kubenswrapper[4979]: E1204 14:31:09.201027 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:31:20 crc kubenswrapper[4979]: I1204 14:31:20.199030 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:31:20 crc kubenswrapper[4979]: E1204 14:31:20.201375 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:31:32 crc kubenswrapper[4979]: I1204 14:31:32.199701 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:31:32 crc kubenswrapper[4979]: E1204 14:31:32.200888 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:31:43 crc kubenswrapper[4979]: I1204 14:31:43.200728 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:31:43 crc kubenswrapper[4979]: E1204 14:31:43.201728 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:31:57 crc kubenswrapper[4979]: I1204 14:31:57.199621 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:31:57 crc kubenswrapper[4979]: E1204 14:31:57.200235 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:32:10 crc kubenswrapper[4979]: I1204 14:32:10.200024 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:32:10 crc kubenswrapper[4979]: E1204 14:32:10.200880 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:32:25 crc kubenswrapper[4979]: I1204 14:32:25.199946 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:32:25 crc kubenswrapper[4979]: E1204 14:32:25.200723 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:32:39 crc kubenswrapper[4979]: I1204 14:32:39.202492 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:32:39 crc kubenswrapper[4979]: E1204 14:32:39.208097 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:32:50 crc kubenswrapper[4979]: I1204 14:32:50.691217 4979 generic.go:334] "Generic (PLEG): container finished" podID="2e03526b-386c-442c-91fd-d64037912796" containerID="e5f057921af81b6b3c93ba28a1848b347bd3ddad12def2adc767dc9c33764be5" exitCode=0 Dec 04 14:32:50 crc kubenswrapper[4979]: I1204 14:32:50.691335 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lckp8/must-gather-qlk82" event={"ID":"2e03526b-386c-442c-91fd-d64037912796","Type":"ContainerDied","Data":"e5f057921af81b6b3c93ba28a1848b347bd3ddad12def2adc767dc9c33764be5"} Dec 04 14:32:50 crc kubenswrapper[4979]: I1204 14:32:50.693093 4979 scope.go:117] "RemoveContainer" containerID="e5f057921af81b6b3c93ba28a1848b347bd3ddad12def2adc767dc9c33764be5" Dec 04 14:32:51 crc kubenswrapper[4979]: I1204 14:32:51.521258 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lckp8_must-gather-qlk82_2e03526b-386c-442c-91fd-d64037912796/gather/0.log" Dec 04 14:32:53 crc kubenswrapper[4979]: I1204 14:32:53.199329 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:32:53 crc kubenswrapper[4979]: E1204 14:32:53.200281 4979 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-8twmz_openshift-machine-config-operator(d9b48d97-5715-467e-b044-4812d6dfea63)\"" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.401540 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lckp8/must-gather-qlk82"] Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.402231 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lckp8/must-gather-qlk82" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="copy" containerID="cri-o://5b366c969068d523774e577730fa69e4cc6a04e844cebae33ccc4dd1def7a044" gracePeriod=2 Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.424764 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lckp8/must-gather-qlk82"] Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.798855 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lckp8_must-gather-qlk82_2e03526b-386c-442c-91fd-d64037912796/copy/0.log" Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.799579 4979 generic.go:334] "Generic (PLEG): container finished" podID="2e03526b-386c-442c-91fd-d64037912796" containerID="5b366c969068d523774e577730fa69e4cc6a04e844cebae33ccc4dd1def7a044" exitCode=143 Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.967905 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lckp8_must-gather-qlk82_2e03526b-386c-442c-91fd-d64037912796/copy/0.log" Dec 04 14:33:00 crc kubenswrapper[4979]: I1204 14:33:00.968504 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.127216 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpd47\" (UniqueName: \"kubernetes.io/projected/2e03526b-386c-442c-91fd-d64037912796-kube-api-access-rpd47\") pod \"2e03526b-386c-442c-91fd-d64037912796\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.127519 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2e03526b-386c-442c-91fd-d64037912796-must-gather-output\") pod \"2e03526b-386c-442c-91fd-d64037912796\" (UID: \"2e03526b-386c-442c-91fd-d64037912796\") " Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.134323 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e03526b-386c-442c-91fd-d64037912796-kube-api-access-rpd47" (OuterVolumeSpecName: "kube-api-access-rpd47") pod "2e03526b-386c-442c-91fd-d64037912796" (UID: "2e03526b-386c-442c-91fd-d64037912796"). InnerVolumeSpecName "kube-api-access-rpd47". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.229928 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpd47\" (UniqueName: \"kubernetes.io/projected/2e03526b-386c-442c-91fd-d64037912796-kube-api-access-rpd47\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.330355 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e03526b-386c-442c-91fd-d64037912796-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2e03526b-386c-442c-91fd-d64037912796" (UID: "2e03526b-386c-442c-91fd-d64037912796"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.332080 4979 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2e03526b-386c-442c-91fd-d64037912796-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.818403 4979 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lckp8_must-gather-qlk82_2e03526b-386c-442c-91fd-d64037912796/copy/0.log" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.821740 4979 scope.go:117] "RemoveContainer" containerID="5b366c969068d523774e577730fa69e4cc6a04e844cebae33ccc4dd1def7a044" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.821937 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lckp8/must-gather-qlk82" Dec 04 14:33:01 crc kubenswrapper[4979]: I1204 14:33:01.847662 4979 scope.go:117] "RemoveContainer" containerID="e5f057921af81b6b3c93ba28a1848b347bd3ddad12def2adc767dc9c33764be5" Dec 04 14:33:02 crc kubenswrapper[4979]: I1204 14:33:02.212387 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e03526b-386c-442c-91fd-d64037912796" path="/var/lib/kubelet/pods/2e03526b-386c-442c-91fd-d64037912796/volumes" Dec 04 14:33:05 crc kubenswrapper[4979]: I1204 14:33:05.199675 4979 scope.go:117] "RemoveContainer" containerID="7100c0c9e287ec6a7791eec6cfc374690389e622490ad41c162c30e6085e49e0" Dec 04 14:33:07 crc kubenswrapper[4979]: I1204 14:33:07.912521 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" event={"ID":"d9b48d97-5715-467e-b044-4812d6dfea63","Type":"ContainerStarted","Data":"43f97dbcb98ed59bd902641ad6ad756cbe76ccd8f8a474d3fb748146434154cc"} Dec 04 14:33:36 crc kubenswrapper[4979]: I1204 14:33:25.940766 4979 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-998648c74-4j4cc" podUID="e729014a-4926-4219-9159-785c4f602117" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.87:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:33:36 crc kubenswrapper[4979]: I1204 14:33:31.775266 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="462e9872-5a8f-4fd0-bd81-cf2ec4471b60" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 04 14:33:36 crc kubenswrapper[4979]: I1204 14:33:33.168542 4979 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-gsffr container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 04 14:33:36 crc kubenswrapper[4979]: I1204 14:33:33.168884 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gsffr" podUID="c2582dcb-a8bc-4c67-8e7f-43c1303112d6" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.739092 4979 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qsm96"] Dec 04 14:34:29 crc kubenswrapper[4979]: E1204 14:34:29.740184 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40a601c-3b93-4693-8bfb-71a87ed29676" containerName="collect-profiles" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.740202 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40a601c-3b93-4693-8bfb-71a87ed29676" containerName="collect-profiles" Dec 04 14:34:29 crc kubenswrapper[4979]: E1204 14:34:29.740241 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="gather" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.740250 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="gather" Dec 04 14:34:29 crc kubenswrapper[4979]: E1204 14:34:29.740469 4979 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="copy" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.740486 4979 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="copy" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.740697 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40a601c-3b93-4693-8bfb-71a87ed29676" containerName="collect-profiles" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.740722 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="gather" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.740738 4979 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e03526b-386c-442c-91fd-d64037912796" containerName="copy" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.742336 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.755139 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qsm96"] Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.891195 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qtwf\" (UniqueName: \"kubernetes.io/projected/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-kube-api-access-5qtwf\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.891267 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-utilities\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.891742 4979 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-catalog-content\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.994281 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qtwf\" (UniqueName: \"kubernetes.io/projected/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-kube-api-access-5qtwf\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.994368 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-utilities\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.994584 4979 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-catalog-content\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.995147 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-catalog-content\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:29 crc kubenswrapper[4979]: I1204 14:34:29.995802 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-utilities\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:30 crc kubenswrapper[4979]: I1204 14:34:30.018901 4979 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qtwf\" (UniqueName: \"kubernetes.io/projected/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-kube-api-access-5qtwf\") pod \"redhat-operators-qsm96\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:30 crc kubenswrapper[4979]: I1204 14:34:30.080235 4979 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:34:30 crc kubenswrapper[4979]: I1204 14:34:30.563817 4979 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qsm96"] Dec 04 14:34:30 crc kubenswrapper[4979]: I1204 14:34:30.804202 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerStarted","Data":"24c318f15f6b71fed963a72776ff5a44a18317e0f7995902c218dcbde52bbb98"} Dec 04 14:34:34 crc kubenswrapper[4979]: I1204 14:34:34.854050 4979 generic.go:334] "Generic (PLEG): container finished" podID="cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" containerID="77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27" exitCode=0 Dec 04 14:34:34 crc kubenswrapper[4979]: I1204 14:34:34.854110 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerDied","Data":"77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27"} Dec 04 14:34:34 crc kubenswrapper[4979]: I1204 14:34:34.858075 4979 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 04 14:34:37 crc kubenswrapper[4979]: I1204 14:34:37.881965 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerStarted","Data":"961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26"} Dec 04 14:34:39 crc kubenswrapper[4979]: I1204 14:34:39.905546 4979 generic.go:334] "Generic (PLEG): container finished" podID="cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" containerID="961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26" exitCode=0 Dec 04 14:34:39 crc kubenswrapper[4979]: I1204 14:34:39.905659 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerDied","Data":"961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26"} Dec 04 14:34:56 crc kubenswrapper[4979]: I1204 14:34:56.776428 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="462e9872-5a8f-4fd0-bd81-cf2ec4471b60" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 04 14:35:02 crc kubenswrapper[4979]: I1204 14:35:02.175676 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerStarted","Data":"5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff"} Dec 04 14:35:02 crc kubenswrapper[4979]: I1204 14:35:02.202710 4979 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qsm96" podStartSLOduration=7.385832129 podStartE2EDuration="33.202687399s" podCreationTimestamp="2025-12-04 14:34:29 +0000 UTC" firstStartedPulling="2025-12-04 14:34:34.857739633 +0000 UTC m=+10299.132035437" lastFinishedPulling="2025-12-04 14:35:00.674594903 +0000 UTC m=+10324.948890707" observedRunningTime="2025-12-04 14:35:02.19858043 +0000 UTC m=+10326.472876254" watchObservedRunningTime="2025-12-04 14:35:02.202687399 +0000 UTC m=+10326.476983193" Dec 04 14:35:10 crc kubenswrapper[4979]: I1204 14:35:10.081294 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:35:10 crc kubenswrapper[4979]: I1204 14:35:10.084989 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:35:10 crc kubenswrapper[4979]: I1204 14:35:10.134584 4979 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:35:10 crc kubenswrapper[4979]: I1204 14:35:10.305564 4979 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:35:10 crc kubenswrapper[4979]: I1204 14:35:10.390356 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qsm96"] Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.274014 4979 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qsm96" podUID="cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" containerName="registry-server" containerID="cri-o://5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff" gracePeriod=2 Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.811055 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.948031 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qtwf\" (UniqueName: \"kubernetes.io/projected/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-kube-api-access-5qtwf\") pod \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.948382 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-catalog-content\") pod \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.948493 4979 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-utilities\") pod \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\" (UID: \"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff\") " Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.949121 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-utilities" (OuterVolumeSpecName: "utilities") pod "cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" (UID: "cb1ecc8c-0f2d-4f88-b9da-89d343d79eff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.949260 4979 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-utilities\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:12 crc kubenswrapper[4979]: I1204 14:35:12.954319 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-kube-api-access-5qtwf" (OuterVolumeSpecName: "kube-api-access-5qtwf") pod "cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" (UID: "cb1ecc8c-0f2d-4f88-b9da-89d343d79eff"). InnerVolumeSpecName "kube-api-access-5qtwf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.052027 4979 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qtwf\" (UniqueName: \"kubernetes.io/projected/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-kube-api-access-5qtwf\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.063868 4979 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" (UID: "cb1ecc8c-0f2d-4f88-b9da-89d343d79eff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.154551 4979 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.283631 4979 generic.go:334] "Generic (PLEG): container finished" podID="cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" containerID="5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff" exitCode=0 Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.283683 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerDied","Data":"5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff"} Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.283714 4979 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qsm96" event={"ID":"cb1ecc8c-0f2d-4f88-b9da-89d343d79eff","Type":"ContainerDied","Data":"24c318f15f6b71fed963a72776ff5a44a18317e0f7995902c218dcbde52bbb98"} Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.283731 4979 scope.go:117] "RemoveContainer" containerID="5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.283860 4979 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qsm96" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.331722 4979 scope.go:117] "RemoveContainer" containerID="961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.336187 4979 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qsm96"] Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.347730 4979 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qsm96"] Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.366007 4979 scope.go:117] "RemoveContainer" containerID="77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.411484 4979 scope.go:117] "RemoveContainer" containerID="5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff" Dec 04 14:35:13 crc kubenswrapper[4979]: E1204 14:35:13.411976 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff\": container with ID starting with 5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff not found: ID does not exist" containerID="5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.412020 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff"} err="failed to get container status \"5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff\": rpc error: code = NotFound desc = could not find container \"5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff\": container with ID starting with 5543129d1efc5727a4e426b748531072b5777128bdcfb68e6d421ded2baac6ff not found: ID does not exist" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.412048 4979 scope.go:117] "RemoveContainer" containerID="961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26" Dec 04 14:35:13 crc kubenswrapper[4979]: E1204 14:35:13.412463 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26\": container with ID starting with 961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26 not found: ID does not exist" containerID="961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.412487 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26"} err="failed to get container status \"961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26\": rpc error: code = NotFound desc = could not find container \"961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26\": container with ID starting with 961f3c9cdfee91c5718dedb2193cce02c7d247b296fcb594c7f77cae9cca7a26 not found: ID does not exist" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.412504 4979 scope.go:117] "RemoveContainer" containerID="77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27" Dec 04 14:35:13 crc kubenswrapper[4979]: E1204 14:35:13.412925 4979 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27\": container with ID starting with 77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27 not found: ID does not exist" containerID="77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27" Dec 04 14:35:13 crc kubenswrapper[4979]: I1204 14:35:13.412951 4979 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27"} err="failed to get container status \"77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27\": rpc error: code = NotFound desc = could not find container \"77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27\": container with ID starting with 77da205bc0b187c133c886b6f31b62ac9c2efad8066b328d8984f203abb50b27 not found: ID does not exist" Dec 04 14:35:14 crc kubenswrapper[4979]: I1204 14:35:14.212360 4979 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb1ecc8c-0f2d-4f88-b9da-89d343d79eff" path="/var/lib/kubelet/pods/cb1ecc8c-0f2d-4f88-b9da-89d343d79eff/volumes" Dec 04 14:35:28 crc kubenswrapper[4979]: I1204 14:35:28.041013 4979 patch_prober.go:28] interesting pod/machine-config-daemon-8twmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 04 14:35:28 crc kubenswrapper[4979]: I1204 14:35:28.041650 4979 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-8twmz" podUID="d9b48d97-5715-467e-b044-4812d6dfea63" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114316105024442 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114316106017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114271350016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114271350015454 5ustar corecore